var/home/core/zuul-output/0000755000175000017500000000000015111015722014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111042350015461 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006347575115111042340017706 0ustar rootrootNov 24 08:52:18 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 08:52:18 crc restorecon[4743]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:52:18 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:19 crc restorecon[4743]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:52:19 crc restorecon[4743]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 08:52:20 crc kubenswrapper[4944]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 08:52:20 crc kubenswrapper[4944]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 08:52:20 crc kubenswrapper[4944]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 08:52:20 crc kubenswrapper[4944]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 08:52:20 crc kubenswrapper[4944]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 08:52:20 crc kubenswrapper[4944]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.005164 4944 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010710 4944 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010751 4944 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010774 4944 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010786 4944 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010796 4944 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010807 4944 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010819 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010832 4944 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010844 4944 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010854 4944 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010865 4944 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010874 4944 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010885 4944 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010895 4944 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010905 4944 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010914 4944 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010924 4944 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010934 4944 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010945 4944 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010955 4944 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010964 4944 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010974 4944 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010985 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.010994 4944 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011028 4944 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011040 4944 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011086 4944 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011096 4944 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011107 4944 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011117 4944 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011126 4944 feature_gate.go:330] unrecognized feature gate: Example Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011135 4944 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011145 4944 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011157 4944 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011167 4944 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011176 4944 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011186 4944 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011197 4944 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011207 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011217 4944 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011226 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011236 4944 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011245 4944 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011256 4944 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011267 4944 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011283 4944 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011296 4944 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011312 4944 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011326 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011340 4944 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011350 4944 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011360 4944 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011371 4944 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011382 4944 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011393 4944 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011404 4944 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011420 4944 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011434 4944 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011446 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011458 4944 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011469 4944 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011479 4944 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011491 4944 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011501 4944 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011513 4944 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011527 4944 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011539 4944 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011549 4944 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011559 4944 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011570 4944 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.011580 4944 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.011826 4944 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.011855 4944 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.011876 4944 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.011891 4944 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.011908 4944 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.011920 4944 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.011936 4944 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.011951 4944 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.011963 4944 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.011976 4944 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.011989 4944 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012002 4944 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012013 4944 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012027 4944 flags.go:64] FLAG: --cgroup-root="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012040 4944 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012083 4944 flags.go:64] FLAG: --client-ca-file="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012095 4944 flags.go:64] FLAG: --cloud-config="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012107 4944 flags.go:64] FLAG: --cloud-provider="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012118 4944 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012134 4944 flags.go:64] FLAG: --cluster-domain="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012145 4944 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012158 4944 flags.go:64] FLAG: --config-dir="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012169 4944 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012182 4944 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012199 4944 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012211 4944 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012223 4944 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012236 4944 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012248 4944 flags.go:64] FLAG: --contention-profiling="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012260 4944 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012272 4944 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012284 4944 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012296 4944 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012311 4944 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012323 4944 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012335 4944 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012347 4944 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012361 4944 flags.go:64] FLAG: --enable-server="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012374 4944 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012390 4944 flags.go:64] FLAG: --event-burst="100" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012403 4944 flags.go:64] FLAG: --event-qps="50" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012414 4944 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012426 4944 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012438 4944 flags.go:64] FLAG: --eviction-hard="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012452 4944 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012479 4944 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012491 4944 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012503 4944 flags.go:64] FLAG: --eviction-soft="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012516 4944 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012528 4944 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012539 4944 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012551 4944 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012563 4944 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012575 4944 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012587 4944 flags.go:64] FLAG: --feature-gates="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012602 4944 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012614 4944 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012627 4944 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012640 4944 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012652 4944 flags.go:64] FLAG: --healthz-port="10248" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012665 4944 flags.go:64] FLAG: --help="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012678 4944 flags.go:64] FLAG: --hostname-override="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012689 4944 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012701 4944 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012712 4944 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012724 4944 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012736 4944 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012747 4944 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012758 4944 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012769 4944 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012781 4944 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012794 4944 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012807 4944 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012822 4944 flags.go:64] FLAG: --kube-reserved="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012834 4944 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012845 4944 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012858 4944 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012876 4944 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012887 4944 flags.go:64] FLAG: --lock-file="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012898 4944 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012912 4944 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012924 4944 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012945 4944 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012956 4944 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012968 4944 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012979 4944 flags.go:64] FLAG: --logging-format="text" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.012992 4944 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013006 4944 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013018 4944 flags.go:64] FLAG: --manifest-url="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013029 4944 flags.go:64] FLAG: --manifest-url-header="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013075 4944 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013089 4944 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013105 4944 flags.go:64] FLAG: --max-pods="110" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013117 4944 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013129 4944 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013141 4944 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013152 4944 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013164 4944 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013178 4944 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013190 4944 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013242 4944 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013255 4944 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013267 4944 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013281 4944 flags.go:64] FLAG: --pod-cidr="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013293 4944 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013313 4944 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013324 4944 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013337 4944 flags.go:64] FLAG: --pods-per-core="0" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013349 4944 flags.go:64] FLAG: --port="10250" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013371 4944 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013382 4944 flags.go:64] FLAG: --provider-id="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013394 4944 flags.go:64] FLAG: --qos-reserved="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013405 4944 flags.go:64] FLAG: --read-only-port="10255" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013420 4944 flags.go:64] FLAG: --register-node="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013432 4944 flags.go:64] FLAG: --register-schedulable="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013444 4944 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013466 4944 flags.go:64] FLAG: --registry-burst="10" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013477 4944 flags.go:64] FLAG: --registry-qps="5" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013489 4944 flags.go:64] FLAG: --reserved-cpus="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013501 4944 flags.go:64] FLAG: --reserved-memory="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013516 4944 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013529 4944 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013540 4944 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013552 4944 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013564 4944 flags.go:64] FLAG: --runonce="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013576 4944 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013588 4944 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013600 4944 flags.go:64] FLAG: --seccomp-default="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013611 4944 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013623 4944 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013636 4944 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013648 4944 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013659 4944 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013670 4944 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013683 4944 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013694 4944 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013706 4944 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013718 4944 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013730 4944 flags.go:64] FLAG: --system-cgroups="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013742 4944 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013760 4944 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013777 4944 flags.go:64] FLAG: --tls-cert-file="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013788 4944 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013804 4944 flags.go:64] FLAG: --tls-min-version="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013815 4944 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013829 4944 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013841 4944 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013853 4944 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013865 4944 flags.go:64] FLAG: --v="2" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013882 4944 flags.go:64] FLAG: --version="false" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013898 4944 flags.go:64] FLAG: --vmodule="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013912 4944 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.013924 4944 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014237 4944 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014255 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014267 4944 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014279 4944 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014291 4944 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014302 4944 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014312 4944 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014323 4944 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014333 4944 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014344 4944 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014354 4944 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014365 4944 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014375 4944 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014386 4944 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014396 4944 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014410 4944 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014423 4944 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014434 4944 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014448 4944 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014461 4944 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014478 4944 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014491 4944 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014504 4944 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014514 4944 feature_gate.go:330] unrecognized feature gate: Example Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014524 4944 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014535 4944 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014545 4944 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014555 4944 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014569 4944 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014579 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014589 4944 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014599 4944 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014609 4944 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014620 4944 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014630 4944 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014640 4944 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014650 4944 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014661 4944 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014672 4944 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014682 4944 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014692 4944 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014702 4944 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014712 4944 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014724 4944 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014738 4944 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014751 4944 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014763 4944 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014774 4944 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014786 4944 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014798 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014814 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014828 4944 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014845 4944 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014856 4944 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014868 4944 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014879 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014889 4944 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014899 4944 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014909 4944 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014920 4944 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014930 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014941 4944 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014951 4944 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014961 4944 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014973 4944 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014984 4944 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.014995 4944 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.015005 4944 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.015015 4944 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.015025 4944 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.015035 4944 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.015168 4944 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.025884 4944 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.025910 4944 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.025978 4944 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.025987 4944 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.025992 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.025997 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026001 4944 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026005 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026009 4944 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026013 4944 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026017 4944 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026021 4944 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026025 4944 feature_gate.go:330] unrecognized feature gate: Example Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026028 4944 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026032 4944 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026037 4944 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026056 4944 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026061 4944 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026067 4944 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026074 4944 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026079 4944 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026083 4944 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026088 4944 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026093 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026097 4944 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026101 4944 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026105 4944 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026109 4944 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026114 4944 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026118 4944 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026122 4944 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026126 4944 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026132 4944 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026137 4944 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026143 4944 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026148 4944 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026155 4944 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026160 4944 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026165 4944 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026169 4944 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026174 4944 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026178 4944 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026182 4944 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026187 4944 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026190 4944 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026194 4944 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026198 4944 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026202 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026206 4944 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026212 4944 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026216 4944 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026220 4944 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026224 4944 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026228 4944 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026232 4944 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026237 4944 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026241 4944 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026246 4944 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026250 4944 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026254 4944 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026257 4944 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026262 4944 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026266 4944 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026270 4944 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026274 4944 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026278 4944 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026282 4944 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026286 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026291 4944 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026295 4944 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026299 4944 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026304 4944 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026309 4944 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.026316 4944 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026447 4944 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026455 4944 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026460 4944 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026464 4944 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026469 4944 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026475 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.026479 4944 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027844 4944 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027856 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027863 4944 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027867 4944 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027873 4944 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027877 4944 feature_gate.go:330] unrecognized feature gate: Example Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027882 4944 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027886 4944 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027890 4944 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027894 4944 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027897 4944 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027901 4944 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027905 4944 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027908 4944 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027912 4944 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027916 4944 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027919 4944 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027923 4944 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027927 4944 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027932 4944 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027937 4944 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027941 4944 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027945 4944 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027949 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027953 4944 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027957 4944 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027960 4944 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027965 4944 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027969 4944 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027972 4944 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027976 4944 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027980 4944 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027984 4944 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027989 4944 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027993 4944 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.027997 4944 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028001 4944 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028006 4944 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028010 4944 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028014 4944 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028018 4944 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028022 4944 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028026 4944 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028030 4944 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028034 4944 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028039 4944 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028059 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028064 4944 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028069 4944 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028073 4944 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028077 4944 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028081 4944 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028084 4944 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028088 4944 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028092 4944 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028097 4944 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028102 4944 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028106 4944 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028109 4944 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028113 4944 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028117 4944 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028121 4944 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028124 4944 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.028129 4944 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.028135 4944 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.029739 4944 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.034584 4944 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.034691 4944 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.036245 4944 server.go:997] "Starting client certificate rotation" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.036268 4944 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.036458 4944 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-07 20:53:39.235896562 +0000 UTC Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.036589 4944 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1068h1m19.199309985s for next certificate rotation Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.074861 4944 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.079669 4944 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.100847 4944 log.go:25] "Validated CRI v1 runtime API" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.140809 4944 log.go:25] "Validated CRI v1 image API" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.144093 4944 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.153604 4944 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-08-47-15-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.153677 4944 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.185409 4944 manager.go:217] Machine: {Timestamp:2025-11-24 08:52:20.180654411 +0000 UTC m=+0.715094953 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:095e1c26-41b9-4bd1-9ffb-7095139764e0 BootID:2763b4ac-9259-4476-9352-1d70e540eb24 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:dc:f0:0e Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:dc:f0:0e Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:e2:78:b6 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:cb:8d:ff Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:5f:2c:0d Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:62:db:cb Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:e5:ce:f8 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:e6:72:e5:41:6a:46 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:2a:3b:d1:53:b9:16 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.186044 4944 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.186686 4944 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.190778 4944 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.191177 4944 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.191230 4944 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.191559 4944 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.191578 4944 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.192190 4944 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.192246 4944 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.192539 4944 state_mem.go:36] "Initialized new in-memory state store" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.192679 4944 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.197595 4944 kubelet.go:418] "Attempting to sync node with API server" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.197636 4944 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.197668 4944 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.197700 4944 kubelet.go:324] "Adding apiserver pod source" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.197726 4944 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.203579 4944 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.204915 4944 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.206912 4944 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.207063 4944 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.15:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.207019 4944 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.207184 4944 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.15:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.207210 4944 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.213986 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.214092 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.214108 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.214122 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.214148 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.214165 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.214181 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.214203 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.214263 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.214284 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.214312 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.214332 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.215102 4944 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.216113 4944 server.go:1280] "Started kubelet" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.218302 4944 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.218606 4944 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.218614 4944 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 08:52:20 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.220099 4944 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.220163 4944 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.220232 4944 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.220250 4944 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.220230 4944 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 12:51:32.407271817 +0000 UTC Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.220280 4944 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.220334 4944 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.220738 4944 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.221029 4944 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.221300 4944 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.15:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.221411 4944 factory.go:55] Registering systemd factory Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.221432 4944 factory.go:221] Registration of the systemd container factory successfully Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.222103 4944 factory.go:153] Registering CRI-O factory Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.222127 4944 factory.go:221] Registration of the crio container factory successfully Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.222256 4944 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.222301 4944 factory.go:103] Registering Raw factory Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.222327 4944 manager.go:1196] Started watching for new ooms in manager Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.223293 4944 server.go:460] "Adding debug handlers to kubelet server" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.224277 4944 manager.go:319] Starting recovery of all containers Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.225149 4944 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.15:6443: connect: connection refused" interval="200ms" Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.225746 4944 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.15:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187ae551057aa7d0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 08:52:20.216006608 +0000 UTC m=+0.750447110,LastTimestamp:2025-11-24 08:52:20.216006608 +0000 UTC m=+0.750447110,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.238839 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.238917 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.238932 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.238953 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.238968 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.238981 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.238998 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.239011 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.239035 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.239123 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.239159 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246403 4944 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246517 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246549 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246572 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246618 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246641 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246660 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246675 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246696 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246712 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246726 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246747 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246839 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246856 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.246878 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247026 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247084 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247131 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247169 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247181 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247216 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247240 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247269 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247283 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247300 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247426 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247448 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247470 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247570 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247661 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247694 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247705 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247742 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247821 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247844 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.247862 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248004 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248106 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248177 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248211 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248234 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248265 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248344 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248370 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248417 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248460 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248485 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248511 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248533 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248628 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.248695 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251202 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251260 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251276 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251289 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251301 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251315 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251329 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251342 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251354 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251365 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251377 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251414 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251426 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251437 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251450 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251464 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251477 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251491 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251502 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251522 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251535 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251551 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251563 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251576 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251593 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251607 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251623 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251639 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251653 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251667 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251680 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251694 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251710 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251723 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251736 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251750 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251763 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251778 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251792 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251805 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251818 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251832 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251845 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251874 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251896 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251930 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251949 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251967 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.251985 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252006 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252022 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252040 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252090 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252116 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252132 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252147 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252158 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252172 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252185 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252206 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252222 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252238 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252250 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252263 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252275 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252289 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252302 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252315 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252329 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252343 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252356 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252371 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252385 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252420 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252437 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252451 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252483 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252496 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252512 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252532 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252545 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252564 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252577 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252589 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252622 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252635 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252649 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252664 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252677 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252708 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252734 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252753 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252771 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252810 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252824 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252838 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252851 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252865 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252876 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252889 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252903 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252915 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252928 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252957 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252977 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.252991 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253004 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253017 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253030 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253062 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253088 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253103 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253118 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253132 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253145 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253157 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253170 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253187 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253199 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253211 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253225 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253262 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253273 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253286 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253298 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253310 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253322 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253334 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253347 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253358 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253370 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253383 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253395 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253408 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253420 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253432 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253445 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253458 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253470 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253481 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253501 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253514 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253527 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253538 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253579 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253592 4944 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253604 4944 reconstruct.go:97] "Volume reconstruction finished" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.253613 4944 reconciler.go:26] "Reconciler: start to sync state" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.257948 4944 manager.go:324] Recovery completed Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.273435 4944 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.275437 4944 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.275506 4944 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.275548 4944 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.275611 4944 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.277392 4944 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.277492 4944 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.15:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.277950 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.279654 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.279697 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.279714 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.280808 4944 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.280837 4944 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.280868 4944 state_mem.go:36] "Initialized new in-memory state store" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.297563 4944 policy_none.go:49] "None policy: Start" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.298896 4944 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.298943 4944 state_mem.go:35] "Initializing new in-memory state store" Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.320600 4944 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.349928 4944 manager.go:334] "Starting Device Plugin manager" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.350309 4944 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.350332 4944 server.go:79] "Starting device plugin registration server" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.350833 4944 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.350875 4944 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.351229 4944 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.351327 4944 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.351338 4944 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.361418 4944 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.376205 4944 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.377100 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.379312 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.379377 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.379396 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.379644 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.379833 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.379895 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.381200 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.381263 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.381277 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.381387 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.381429 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.381439 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.381686 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.381929 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.381989 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.383013 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.383087 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.383107 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.383018 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.383159 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.383173 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.383333 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.383351 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.383360 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.384034 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.384100 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.384119 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.384330 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.384438 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.384480 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.384680 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.384716 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.384732 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.385202 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.385231 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.385243 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.385416 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.385442 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.385457 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.385805 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.385879 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.386886 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.386915 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.386931 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.426388 4944 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.15:6443: connect: connection refused" interval="400ms" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.451090 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.452898 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.452945 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.452959 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.452997 4944 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.453755 4944 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.15:6443: connect: connection refused" node="crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.455831 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.455866 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.455890 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.455910 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.455927 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.455999 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.456099 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.456142 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.456160 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.456185 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.456205 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.456222 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.456238 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.456275 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.456302 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558103 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558179 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558199 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558217 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558232 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558249 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558268 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558288 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558306 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558323 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558346 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558369 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558347 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558405 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558439 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558388 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558467 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558508 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558535 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558542 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558541 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558532 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558586 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558571 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558571 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558616 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558557 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558591 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558555 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.558624 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.654894 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.656525 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.656590 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.656603 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.656644 4944 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.657372 4944 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.15:6443: connect: connection refused" node="crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.706844 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.713669 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.731592 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.738620 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: I1124 08:52:20.747408 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.754476 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-390e8cc246284946a5c8046a3fba99747f8a7ef7e6d9e92819c1f6c29ae40c39 WatchSource:0}: Error finding container 390e8cc246284946a5c8046a3fba99747f8a7ef7e6d9e92819c1f6c29ae40c39: Status 404 returned error can't find the container with id 390e8cc246284946a5c8046a3fba99747f8a7ef7e6d9e92819c1f6c29ae40c39 Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.755183 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-24b66731f1b0ed9e665119977dd28b8344418d60b4e951ff3e95775525a71ce5 WatchSource:0}: Error finding container 24b66731f1b0ed9e665119977dd28b8344418d60b4e951ff3e95775525a71ce5: Status 404 returned error can't find the container with id 24b66731f1b0ed9e665119977dd28b8344418d60b4e951ff3e95775525a71ce5 Nov 24 08:52:20 crc kubenswrapper[4944]: W1124 08:52:20.775072 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-6568b9a96fe7960043425f4bd7adebd23bf84832d6e080282d4bcf1d4e84d95e WatchSource:0}: Error finding container 6568b9a96fe7960043425f4bd7adebd23bf84832d6e080282d4bcf1d4e84d95e: Status 404 returned error can't find the container with id 6568b9a96fe7960043425f4bd7adebd23bf84832d6e080282d4bcf1d4e84d95e Nov 24 08:52:20 crc kubenswrapper[4944]: E1124 08:52:20.828296 4944 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.15:6443: connect: connection refused" interval="800ms" Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.057775 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.059630 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.059684 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.059693 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.059725 4944 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:52:21 crc kubenswrapper[4944]: E1124 08:52:21.060370 4944 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.15:6443: connect: connection refused" node="crc" Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.219865 4944 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.220869 4944 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 04:30:38.099273438 +0000 UTC Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.220944 4944 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 955h38m16.878332837s for next certificate rotation Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.281700 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"390e8cc246284946a5c8046a3fba99747f8a7ef7e6d9e92819c1f6c29ae40c39"} Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.283639 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6568b9a96fe7960043425f4bd7adebd23bf84832d6e080282d4bcf1d4e84d95e"} Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.284867 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"30695981ea98319e7996aee8f60c2b483f995c5aed8b033138d65bd904ac809a"} Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.286235 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e065067b1fdeea82873b05e26b63a065d8d58fdacc9f9819e9d59ce23d5c0304"} Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.287822 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"24b66731f1b0ed9e665119977dd28b8344418d60b4e951ff3e95775525a71ce5"} Nov 24 08:52:21 crc kubenswrapper[4944]: W1124 08:52:21.304870 4944 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:21 crc kubenswrapper[4944]: E1124 08:52:21.304972 4944 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.15:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:52:21 crc kubenswrapper[4944]: W1124 08:52:21.309824 4944 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:21 crc kubenswrapper[4944]: E1124 08:52:21.309945 4944 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.15:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:52:21 crc kubenswrapper[4944]: W1124 08:52:21.607023 4944 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:21 crc kubenswrapper[4944]: E1124 08:52:21.607159 4944 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.15:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:52:21 crc kubenswrapper[4944]: E1124 08:52:21.629666 4944 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.15:6443: connect: connection refused" interval="1.6s" Nov 24 08:52:21 crc kubenswrapper[4944]: W1124 08:52:21.804000 4944 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:21 crc kubenswrapper[4944]: E1124 08:52:21.804124 4944 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.15:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.861258 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.863292 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.863347 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.863359 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:21 crc kubenswrapper[4944]: I1124 08:52:21.863389 4944 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:52:21 crc kubenswrapper[4944]: E1124 08:52:21.864041 4944 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.15:6443: connect: connection refused" node="crc" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.219890 4944 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.293876 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd"} Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.293939 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61"} Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.293950 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609"} Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.293959 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1"} Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.294097 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.295289 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.295344 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.295364 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.296171 4944 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920" exitCode=0 Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.296267 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920"} Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.296303 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.297416 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.297461 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.297473 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.297786 4944 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="af2547b045977dfa45ebb654aa19b3680783b92e2fc4c446cc7773a1cc367ac0" exitCode=0 Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.297854 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"af2547b045977dfa45ebb654aa19b3680783b92e2fc4c446cc7773a1cc367ac0"} Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.297952 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.299037 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.299110 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.299130 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.299265 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.299671 4944 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364" exitCode=0 Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.299756 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364"} Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.299771 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.300832 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.300872 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.300839 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.300889 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.300908 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.300927 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.302902 4944 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b" exitCode=0 Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.302944 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b"} Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.303254 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.304186 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.304221 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:22 crc kubenswrapper[4944]: I1124 08:52:22.304235 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.219242 4944 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:23 crc kubenswrapper[4944]: E1124 08:52:23.230973 4944 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.15:6443: connect: connection refused" interval="3.2s" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.308674 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb"} Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.308727 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1"} Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.308739 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611"} Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.308751 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38"} Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.310815 4944 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a6ca92ef92ca84a95fb520801e4fd4ea48d6c931b0ddc30cd02e53ef0b6844e6" exitCode=0 Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.310894 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a6ca92ef92ca84a95fb520801e4fd4ea48d6c931b0ddc30cd02e53ef0b6844e6"} Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.310943 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.312275 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.312311 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.312326 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.312601 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ae6958eb78d429d6c25449d6b5595b7ef19937a8db223b185317ae344b8dd947"} Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.312643 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.313552 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.313597 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.313636 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.317597 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.317597 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"94d33673d9208d6570c2853f6644328f2997815a8bfd48a5d7338d12bab2f77d"} Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.317671 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c10a53461679aed9575532a08cb82b6bae353bf170c7a663369aeb4dac512e06"} Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.317692 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.317692 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"739cd0cec0794b22aa306490236c6e42aef2aab4b8787cfed8752e727d1c48c5"} Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.318361 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.318388 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.318398 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.319146 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.319181 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.319193 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:23 crc kubenswrapper[4944]: W1124 08:52:23.365805 4944 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:23 crc kubenswrapper[4944]: E1124 08:52:23.365911 4944 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.15:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.464877 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.468371 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.468433 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.468449 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:23 crc kubenswrapper[4944]: I1124 08:52:23.468486 4944 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:52:23 crc kubenswrapper[4944]: E1124 08:52:23.469083 4944 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.15:6443: connect: connection refused" node="crc" Nov 24 08:52:23 crc kubenswrapper[4944]: W1124 08:52:23.556739 4944 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.15:6443: connect: connection refused Nov 24 08:52:23 crc kubenswrapper[4944]: E1124 08:52:23.556877 4944 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.15:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.322684 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64"} Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.322836 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.324472 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.324511 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.324526 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.325587 4944 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f17de1645896f406cbb2127db6b18e1d5878427fb97071da3b7287b374e9f68d" exitCode=0 Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.325675 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.326254 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.326641 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f17de1645896f406cbb2127db6b18e1d5878427fb97071da3b7287b374e9f68d"} Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.326733 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.326757 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.328304 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.328335 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.328349 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.328941 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.328972 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.328983 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.329567 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.329592 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.329604 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:24 crc kubenswrapper[4944]: I1124 08:52:24.779584 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:25 crc kubenswrapper[4944]: I1124 08:52:25.331854 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:25 crc kubenswrapper[4944]: I1124 08:52:25.333246 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7e4a6b4311728134d7738bfff9ec09c97eba333399a4ce0bf9415c1e890f595f"} Nov 24 08:52:25 crc kubenswrapper[4944]: I1124 08:52:25.333288 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fe718044da2e37afec38a48ac8406c191567c30643a1543f388465850db7b61b"} Nov 24 08:52:25 crc kubenswrapper[4944]: I1124 08:52:25.333300 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"46f5667173b9b85fc722c3406b28ab925849205291404cd7458f74a325b53977"} Nov 24 08:52:25 crc kubenswrapper[4944]: I1124 08:52:25.333615 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:25 crc kubenswrapper[4944]: I1124 08:52:25.333640 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:25 crc kubenswrapper[4944]: I1124 08:52:25.333648 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.298223 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.298568 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.300389 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.300451 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.300555 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.340570 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e3d0ded48080da67f4eea4447b38ce9428173a328ac0ac59fdf8eb6def95695b"} Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.340648 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"090878d55ac29d8999925373bb5c7ff1cd2130194803c25589c17f05875ceb8b"} Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.340659 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.340693 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.341907 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.341950 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.341962 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.341917 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.341992 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.342000 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.461847 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.669179 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.671189 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.671230 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.671240 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:26 crc kubenswrapper[4944]: I1124 08:52:26.671265 4944 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:52:27 crc kubenswrapper[4944]: I1124 08:52:27.226270 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:27 crc kubenswrapper[4944]: I1124 08:52:27.343580 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:27 crc kubenswrapper[4944]: I1124 08:52:27.343580 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:27 crc kubenswrapper[4944]: I1124 08:52:27.345241 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:27 crc kubenswrapper[4944]: I1124 08:52:27.345284 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:27 crc kubenswrapper[4944]: I1124 08:52:27.345298 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:27 crc kubenswrapper[4944]: I1124 08:52:27.346175 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:27 crc kubenswrapper[4944]: I1124 08:52:27.346251 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:27 crc kubenswrapper[4944]: I1124 08:52:27.346265 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:27 crc kubenswrapper[4944]: I1124 08:52:27.364243 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.345993 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.345993 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.347272 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.347350 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.347369 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.347474 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.347540 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.347551 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.545849 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.546112 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.548062 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.548104 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:28 crc kubenswrapper[4944]: I1124 08:52:28.548117 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:29 crc kubenswrapper[4944]: I1124 08:52:29.459132 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:29 crc kubenswrapper[4944]: I1124 08:52:29.459376 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:29 crc kubenswrapper[4944]: I1124 08:52:29.460853 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:29 crc kubenswrapper[4944]: I1124 08:52:29.460900 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:29 crc kubenswrapper[4944]: I1124 08:52:29.460912 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:29 crc kubenswrapper[4944]: I1124 08:52:29.465325 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:30 crc kubenswrapper[4944]: I1124 08:52:30.136507 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:30 crc kubenswrapper[4944]: I1124 08:52:30.351906 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:30 crc kubenswrapper[4944]: I1124 08:52:30.353321 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:30 crc kubenswrapper[4944]: I1124 08:52:30.353390 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:30 crc kubenswrapper[4944]: I1124 08:52:30.353410 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:30 crc kubenswrapper[4944]: E1124 08:52:30.361637 4944 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 08:52:31 crc kubenswrapper[4944]: I1124 08:52:31.355440 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:31 crc kubenswrapper[4944]: I1124 08:52:31.356723 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:31 crc kubenswrapper[4944]: I1124 08:52:31.356791 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:31 crc kubenswrapper[4944]: I1124 08:52:31.356806 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:31 crc kubenswrapper[4944]: I1124 08:52:31.360264 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:31 crc kubenswrapper[4944]: I1124 08:52:31.546926 4944 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 08:52:31 crc kubenswrapper[4944]: I1124 08:52:31.547103 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:52:32 crc kubenswrapper[4944]: I1124 08:52:32.198495 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:32 crc kubenswrapper[4944]: I1124 08:52:32.358137 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:32 crc kubenswrapper[4944]: I1124 08:52:32.360105 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:32 crc kubenswrapper[4944]: I1124 08:52:32.360155 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:32 crc kubenswrapper[4944]: I1124 08:52:32.360166 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:33 crc kubenswrapper[4944]: I1124 08:52:33.361696 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:33 crc kubenswrapper[4944]: I1124 08:52:33.362811 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:33 crc kubenswrapper[4944]: I1124 08:52:33.362872 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:33 crc kubenswrapper[4944]: I1124 08:52:33.362888 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:33 crc kubenswrapper[4944]: I1124 08:52:33.998717 4944 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 08:52:33 crc kubenswrapper[4944]: I1124 08:52:33.998781 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.023652 4944 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.023751 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.249682 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.249964 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.251459 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.251514 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.251528 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.282640 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.364845 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.366336 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.366410 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.366425 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:34 crc kubenswrapper[4944]: I1124 08:52:34.380022 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 08:52:35 crc kubenswrapper[4944]: I1124 08:52:35.367841 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:35 crc kubenswrapper[4944]: I1124 08:52:35.369020 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:35 crc kubenswrapper[4944]: I1124 08:52:35.369130 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:35 crc kubenswrapper[4944]: I1124 08:52:35.369148 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:37 crc kubenswrapper[4944]: I1124 08:52:37.232146 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:37 crc kubenswrapper[4944]: I1124 08:52:37.232393 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:37 crc kubenswrapper[4944]: I1124 08:52:37.233748 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:37 crc kubenswrapper[4944]: I1124 08:52:37.233794 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:37 crc kubenswrapper[4944]: I1124 08:52:37.233809 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:37 crc kubenswrapper[4944]: I1124 08:52:37.240162 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:37 crc kubenswrapper[4944]: I1124 08:52:37.374580 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:52:37 crc kubenswrapper[4944]: I1124 08:52:37.374634 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:37 crc kubenswrapper[4944]: I1124 08:52:37.375760 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:37 crc kubenswrapper[4944]: I1124 08:52:37.375793 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:37 crc kubenswrapper[4944]: I1124 08:52:37.375803 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:38 crc kubenswrapper[4944]: E1124 08:52:38.993638 4944 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 24 08:52:38 crc kubenswrapper[4944]: I1124 08:52:38.995966 4944 trace.go:236] Trace[583402413]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 08:52:24.418) (total time: 14577ms): Nov 24 08:52:38 crc kubenswrapper[4944]: Trace[583402413]: ---"Objects listed" error: 14577ms (08:52:38.995) Nov 24 08:52:38 crc kubenswrapper[4944]: Trace[583402413]: [14.577437717s] [14.577437717s] END Nov 24 08:52:38 crc kubenswrapper[4944]: I1124 08:52:38.996007 4944 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 08:52:38 crc kubenswrapper[4944]: I1124 08:52:38.996266 4944 trace.go:236] Trace[345577859]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 08:52:24.348) (total time: 14647ms): Nov 24 08:52:38 crc kubenswrapper[4944]: Trace[345577859]: ---"Objects listed" error: 14647ms (08:52:38.996) Nov 24 08:52:38 crc kubenswrapper[4944]: Trace[345577859]: [14.647626699s] [14.647626699s] END Nov 24 08:52:38 crc kubenswrapper[4944]: I1124 08:52:38.996284 4944 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 08:52:38 crc kubenswrapper[4944]: I1124 08:52:38.997460 4944 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 08:52:38 crc kubenswrapper[4944]: I1124 08:52:38.997473 4944 trace.go:236] Trace[1975985078]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 08:52:28.760) (total time: 10236ms): Nov 24 08:52:38 crc kubenswrapper[4944]: Trace[1975985078]: ---"Objects listed" error: 10236ms (08:52:38.997) Nov 24 08:52:38 crc kubenswrapper[4944]: Trace[1975985078]: [10.236560334s] [10.236560334s] END Nov 24 08:52:38 crc kubenswrapper[4944]: I1124 08:52:38.997670 4944 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 08:52:38 crc kubenswrapper[4944]: I1124 08:52:38.998295 4944 trace.go:236] Trace[889387884]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 08:52:27.320) (total time: 11678ms): Nov 24 08:52:38 crc kubenswrapper[4944]: Trace[889387884]: ---"Objects listed" error: 11677ms (08:52:38.998) Nov 24 08:52:38 crc kubenswrapper[4944]: Trace[889387884]: [11.678068204s] [11.678068204s] END Nov 24 08:52:38 crc kubenswrapper[4944]: I1124 08:52:38.998356 4944 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 08:52:38 crc kubenswrapper[4944]: E1124 08:52:38.998944 4944 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.063073 4944 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48930->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.063186 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48930->192.168.126.11:17697: read: connection reset by peer" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.063719 4944 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.063813 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.091195 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.103822 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.210762 4944 apiserver.go:52] "Watching apiserver" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.214825 4944 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.215109 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.215424 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.215508 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.215686 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.215763 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.215804 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.215834 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.216019 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.216090 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.216227 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.221022 4944 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.229162 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.229919 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.230076 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.230212 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.230313 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.245819 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.245963 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.245968 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.252679 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.263230 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-9zglk"] Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.263645 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9zglk" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.274893 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.277954 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.280428 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.294993 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300136 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300190 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300215 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300235 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300254 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300273 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300294 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300312 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300333 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300351 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300370 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300387 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300408 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300427 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300445 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300462 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300483 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300504 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300524 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300541 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300559 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300574 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300592 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300608 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300697 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300720 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300741 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300762 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300781 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300801 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300808 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300815 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300819 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300889 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300916 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300937 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300956 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300977 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300999 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301018 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301038 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301077 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301097 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301117 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301139 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301161 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301180 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301208 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301228 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301249 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301301 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301323 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301344 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301365 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301384 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301405 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301424 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301443 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301462 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301480 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301498 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301520 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301538 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301558 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301576 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301598 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301618 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301638 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301660 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301678 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301699 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301719 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301740 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301759 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301776 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301792 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301810 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301830 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301847 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301863 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301879 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301895 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301912 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301928 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301945 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301962 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301979 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301997 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302016 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302034 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302116 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302134 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302154 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302175 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302193 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302211 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302227 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.300799 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301019 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301073 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301214 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301239 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301396 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301616 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301681 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301734 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.301950 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302414 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302404 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302447 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302635 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302665 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302709 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302736 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302705 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302817 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302936 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302969 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303030 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303141 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303150 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303193 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303367 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303392 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303439 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303465 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303478 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303578 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303693 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303750 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303764 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303801 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303808 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.303901 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.304010 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.304023 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.304414 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.304422 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.302243 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.304915 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.304998 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305114 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305192 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305263 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305331 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305397 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305467 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305532 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305597 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305663 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305727 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305791 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305857 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305921 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.305988 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306083 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306167 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306234 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306305 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306377 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306447 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306547 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306620 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306696 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306767 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306835 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306904 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306973 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307037 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307142 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307215 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307296 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307374 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307449 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307520 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307587 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307653 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307721 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307791 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307864 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307927 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307993 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.308080 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.308157 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.308236 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.308299 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.308365 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.308433 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.311430 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306620 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306779 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306796 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306840 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.311923 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.306936 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.307250 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.309538 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.309557 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.309838 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.310215 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.312012 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.310287 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.310422 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.310645 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.310737 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.310815 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.310996 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.311133 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.311215 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.311341 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.311714 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.311754 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.311772 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.311986 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.312330 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.312416 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.312565 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.312582 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.312854 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.313015 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.313022 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.313310 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.313694 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.313844 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.313857 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.313899 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.314012 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.314028 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.314148 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.314299 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.314459 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.314480 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.314628 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.314954 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.315072 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.315114 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.314826 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.315164 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.315366 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.315475 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.315628 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.315725 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.315844 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.315902 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.315929 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.316091 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.316314 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.316352 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.316749 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.316832 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.316861 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.317944 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.318517 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.318849 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.319342 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.319561 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.320709 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.321343 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.321369 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.321409 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.321783 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.322231 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.322313 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.322462 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.322862 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.323267 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.323623 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.324072 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.324184 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.312488 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.324733 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.324850 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325574 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325620 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325646 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325667 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325694 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325712 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325731 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325750 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325769 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325786 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325803 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325823 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325852 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325873 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325892 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325920 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325946 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325969 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.325979 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326126 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326152 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326174 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326197 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326223 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326244 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326264 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326259 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326290 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326312 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326336 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326357 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326380 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326400 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326422 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326444 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326463 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326464 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326485 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326489 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326576 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.326601 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:52:39.826570281 +0000 UTC m=+20.361010903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326690 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326720 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326748 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326766 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326778 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326786 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326828 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326836 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326856 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326884 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326909 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326940 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326964 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.326989 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327036 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327076 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327100 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327122 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327144 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327203 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327228 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327254 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327273 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327294 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327321 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327342 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327368 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327404 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327917 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.330146 4944 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327365 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327380 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.327467 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.328091 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.328404 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.328448 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.328850 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.329126 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.329308 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.329515 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.329794 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.329964 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.329994 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.329991 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.330109 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.330320 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.330384 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.330475 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.330422 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.330554 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.331493 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.331547 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.331764 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.332238 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.332373 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.332756 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.333012 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.333234 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.333241 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.333446 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.333264 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.333700 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.333545 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.334494 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.334812 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.335502 4944 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.335559 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.335655 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:39.835629369 +0000 UTC m=+20.370070031 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.335694 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.335906 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336119 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336302 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.336327 4944 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336416 4944 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336426 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.336503 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:39.836485017 +0000 UTC m=+20.370925479 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336782 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336816 4944 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336830 4944 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336844 4944 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336891 4944 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336913 4944 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336927 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336941 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336951 4944 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336960 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336972 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336981 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336990 4944 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337003 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337014 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337024 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337036 4944 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337060 4944 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337070 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337082 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337092 4944 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337102 4944 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337112 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337122 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337133 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337144 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337156 4944 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337166 4944 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337176 4944 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337186 4944 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337198 4944 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337210 4944 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337222 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337235 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337248 4944 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337312 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337322 4944 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337332 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337342 4944 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337351 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337361 4944 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337370 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337381 4944 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337393 4944 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337405 4944 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337418 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337430 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337443 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337455 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337465 4944 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337478 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337489 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337503 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337515 4944 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337526 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337535 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337544 4944 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337553 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337563 4944 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337574 4944 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337585 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337596 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337607 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337618 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337633 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337643 4944 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337656 4944 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337667 4944 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337683 4944 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337694 4944 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337705 4944 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337714 4944 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337723 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.336848 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337706 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337707 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337733 4944 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337824 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337839 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337853 4944 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337864 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337877 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337887 4944 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337899 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337910 4944 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337921 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337932 4944 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337943 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337953 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337964 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337975 4944 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337986 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.337997 4944 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338010 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338024 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338074 4944 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338093 4944 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338107 4944 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338119 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338132 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338146 4944 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338158 4944 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338170 4944 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338184 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338195 4944 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338210 4944 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338221 4944 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338232 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338243 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338253 4944 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338265 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338274 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338284 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338293 4944 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338302 4944 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338315 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338326 4944 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338338 4944 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338349 4944 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338360 4944 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338369 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338377 4944 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338388 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338397 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338406 4944 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338415 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338424 4944 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338456 4944 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338469 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.338482 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.339025 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.339176 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.339266 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.339885 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.342672 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.342772 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.347827 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.348003 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.348275 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.349603 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.351342 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.351576 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.352402 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.352855 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.353103 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.353124 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.353140 4944 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.353360 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:39.853333783 +0000 UTC m=+20.387774245 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.353817 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.354631 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.355409 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.355427 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.355437 4944 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.355488 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:39.855477581 +0000 UTC m=+20.389918043 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.356681 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.356849 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.357009 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.357575 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.357649 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.358829 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.359627 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.359787 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.359870 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.360182 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.360371 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.368489 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.369210 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.376712 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.380994 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.381539 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.384939 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.391165 4944 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64" exitCode=255 Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.391292 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64"} Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.392335 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.393755 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.399604 4944 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.413561 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.433333 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439038 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l95s7\" (UniqueName: \"kubernetes.io/projected/3c792a6e-0110-4e35-923e-c1a38f32e8f9-kube-api-access-l95s7\") pod \"node-resolver-9zglk\" (UID: \"3c792a6e-0110-4e35-923e-c1a38f32e8f9\") " pod="openshift-dns/node-resolver-9zglk" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439120 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439141 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3c792a6e-0110-4e35-923e-c1a38f32e8f9-hosts-file\") pod \"node-resolver-9zglk\" (UID: \"3c792a6e-0110-4e35-923e-c1a38f32e8f9\") " pod="openshift-dns/node-resolver-9zglk" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439180 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439221 4944 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439234 4944 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439245 4944 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439256 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439270 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439281 4944 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439291 4944 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439301 4944 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439312 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439322 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439331 4944 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439411 4944 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439477 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439499 4944 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439512 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439524 4944 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439535 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439538 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439547 4944 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439590 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439590 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439601 4944 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439673 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439699 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439711 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439724 4944 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439736 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439747 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439759 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439771 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439784 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439796 4944 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439808 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439819 4944 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439831 4944 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439843 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439853 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439866 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439880 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439892 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439903 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439914 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439925 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439936 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439947 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439959 4944 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439972 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439983 4944 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.439997 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440010 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440021 4944 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440033 4944 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440064 4944 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440077 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440090 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440103 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440116 4944 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440128 4944 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440140 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440156 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440170 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440183 4944 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440197 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440212 4944 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440260 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.440273 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.446204 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.456513 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.466896 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.477568 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.489955 4944 scope.go:117] "RemoveContainer" containerID="f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.491533 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.493122 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.506733 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.518845 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.528556 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.534491 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.541656 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l95s7\" (UniqueName: \"kubernetes.io/projected/3c792a6e-0110-4e35-923e-c1a38f32e8f9-kube-api-access-l95s7\") pod \"node-resolver-9zglk\" (UID: \"3c792a6e-0110-4e35-923e-c1a38f32e8f9\") " pod="openshift-dns/node-resolver-9zglk" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.541707 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3c792a6e-0110-4e35-923e-c1a38f32e8f9-hosts-file\") pod \"node-resolver-9zglk\" (UID: \"3c792a6e-0110-4e35-923e-c1a38f32e8f9\") " pod="openshift-dns/node-resolver-9zglk" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.541772 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.541824 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3c792a6e-0110-4e35-923e-c1a38f32e8f9-hosts-file\") pod \"node-resolver-9zglk\" (UID: \"3c792a6e-0110-4e35-923e-c1a38f32e8f9\") " pod="openshift-dns/node-resolver-9zglk" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.549000 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.572344 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l95s7\" (UniqueName: \"kubernetes.io/projected/3c792a6e-0110-4e35-923e-c1a38f32e8f9-kube-api-access-l95s7\") pod \"node-resolver-9zglk\" (UID: \"3c792a6e-0110-4e35-923e-c1a38f32e8f9\") " pod="openshift-dns/node-resolver-9zglk" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.573749 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9zglk" Nov 24 08:52:39 crc kubenswrapper[4944]: W1124 08:52:39.573881 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-2d78cdb82b4603ad4f5124d780b9031a527ba0a480a14a9d1c2ce99c90da9f0c WatchSource:0}: Error finding container 2d78cdb82b4603ad4f5124d780b9031a527ba0a480a14a9d1c2ce99c90da9f0c: Status 404 returned error can't find the container with id 2d78cdb82b4603ad4f5124d780b9031a527ba0a480a14a9d1c2ce99c90da9f0c Nov 24 08:52:39 crc kubenswrapper[4944]: W1124 08:52:39.579094 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-04342ddb0833f8a21e7602530ac6e48a02bc16db8a6ed3fb9baf2d6ec4197aa3 WatchSource:0}: Error finding container 04342ddb0833f8a21e7602530ac6e48a02bc16db8a6ed3fb9baf2d6ec4197aa3: Status 404 returned error can't find the container with id 04342ddb0833f8a21e7602530ac6e48a02bc16db8a6ed3fb9baf2d6ec4197aa3 Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.843924 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.843995 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.844025 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.844164 4944 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.844180 4944 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.844227 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:52:40.844188559 +0000 UTC m=+21.378629011 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.844289 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:40.844279302 +0000 UTC m=+21.378719764 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.844313 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:40.844297882 +0000 UTC m=+21.378738344 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.944859 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:39 crc kubenswrapper[4944]: I1124 08:52:39.944926 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.945110 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.945128 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.945187 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.945138 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.945205 4944 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.945223 4944 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.945291 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:40.945265064 +0000 UTC m=+21.479705696 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:39 crc kubenswrapper[4944]: E1124 08:52:39.945338 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:40.945321006 +0000 UTC m=+21.479761468 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.059687 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-dg4dk"] Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.060117 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.063636 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.064397 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.064898 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.065751 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.065837 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.066126 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7q5tz"] Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.072262 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.072471 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-llhq9"] Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.073132 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-tc6n6"] Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.073294 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.074301 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.077579 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.077619 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.077822 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.077957 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.078017 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.078058 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.078146 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.078450 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.078559 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.078639 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.078779 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.078897 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.079089 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.079241 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.111020 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.135458 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146216 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e4e5c60c-e37d-43c8-860e-39ef930adc16-proxy-tls\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146276 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5ff38b2b-032d-47df-8836-105e0cfae835-cni-binary-copy\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146323 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-cnibin\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146351 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtq7z\" (UniqueName: \"kubernetes.io/projected/e4e5c60c-e37d-43c8-860e-39ef930adc16-kube-api-access-dtq7z\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146379 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-multus-cni-dir\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146477 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-os-release\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146525 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-run-k8s-cni-cncf-io\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146554 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e4e5c60c-e37d-43c8-860e-39ef930adc16-rootfs\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146575 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-var-lib-kubelet\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146599 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-multus-conf-dir\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146635 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-run-netns\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146653 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-hostroot\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146688 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-system-cni-dir\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146732 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4e5c60c-e37d-43c8-860e-39ef930adc16-mcd-auth-proxy-config\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146751 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-multus-socket-dir-parent\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146770 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-var-lib-cni-bin\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146792 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-var-lib-cni-multus\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.146833 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5ff38b2b-032d-47df-8836-105e0cfae835-multus-daemon-config\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.151091 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.163642 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.177023 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.190923 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.203555 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.218797 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.229116 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.241754 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.247830 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-system-cni-dir\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248034 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-system-cni-dir\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248073 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-etc-kubernetes\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248159 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-ovn\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248192 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4e5c60c-e37d-43c8-860e-39ef930adc16-mcd-auth-proxy-config\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248215 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5ff38b2b-032d-47df-8836-105e0cfae835-multus-daemon-config\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248233 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-log-socket\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248250 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e4e5c60c-e37d-43c8-860e-39ef930adc16-proxy-tls\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248273 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5ff38b2b-032d-47df-8836-105e0cfae835-cni-binary-copy\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248291 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-script-lib\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248310 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn57m\" (UniqueName: \"kubernetes.io/projected/aeb94161-d41a-420b-b5ed-9b65abbe91a3-kube-api-access-bn57m\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248330 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-systemd\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248352 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248386 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzdlj\" (UniqueName: \"kubernetes.io/projected/c5899104-758a-40da-9612-7334e6616928-kube-api-access-bzdlj\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248413 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-multus-conf-dir\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248434 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-run-multus-certs\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248453 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-system-cni-dir\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.248664 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-multus-conf-dir\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.249124 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4e5c60c-e37d-43c8-860e-39ef930adc16-mcd-auth-proxy-config\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.249441 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5ff38b2b-032d-47df-8836-105e0cfae835-cni-binary-copy\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.249665 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5ff38b2b-032d-47df-8836-105e0cfae835-multus-daemon-config\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.249814 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtq7z\" (UniqueName: \"kubernetes.io/projected/e4e5c60c-e37d-43c8-860e-39ef930adc16-kube-api-access-dtq7z\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.250349 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-os-release\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.250756 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-os-release\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.250825 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-run-k8s-cni-cncf-io\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.250903 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-run-k8s-cni-cncf-io\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.250947 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-node-log\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.250972 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-ovn-kubernetes\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.250999 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovn-node-metrics-cert\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251026 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c5899104-758a-40da-9612-7334e6616928-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251075 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-slash\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251098 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-bin\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251117 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c5899104-758a-40da-9612-7334e6616928-cni-binary-copy\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251140 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-var-lib-cni-bin\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251265 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-var-lib-cni-bin\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251327 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-var-lib-cni-multus\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251344 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-var-lib-cni-multus\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251367 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-kubelet\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251421 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-multus-socket-dir-parent\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251449 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-os-release\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251477 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-etc-openvswitch\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251509 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-multus-socket-dir-parent\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251523 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-var-lib-openvswitch\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251555 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-config\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251616 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-netd\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251701 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-cnibin\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251794 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-systemd-units\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251825 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-netns\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251850 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-openvswitch\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251868 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251888 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-cnibin\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251891 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-cnibin\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251921 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e4e5c60c-e37d-43c8-860e-39ef930adc16-rootfs\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251940 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-multus-cni-dir\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251947 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e4e5c60c-e37d-43c8-860e-39ef930adc16-rootfs\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251959 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-var-lib-kubelet\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251976 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p26h\" (UniqueName: \"kubernetes.io/projected/5ff38b2b-032d-47df-8836-105e0cfae835-kube-api-access-9p26h\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.251997 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-run-netns\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.252004 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-multus-cni-dir\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.252016 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-hostroot\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.252038 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-env-overrides\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.252091 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-hostroot\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.252038 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-var-lib-kubelet\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.252175 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-run-netns\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.253410 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e4e5c60c-e37d-43c8-860e-39ef930adc16-proxy-tls\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.255847 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.264071 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.268061 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtq7z\" (UniqueName: \"kubernetes.io/projected/e4e5c60c-e37d-43c8-860e-39ef930adc16-kube-api-access-dtq7z\") pod \"machine-config-daemon-llhq9\" (UID: \"e4e5c60c-e37d-43c8-860e-39ef930adc16\") " pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.276678 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.288519 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.289139 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.289996 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.291311 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.291937 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.293085 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.293698 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.294374 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.295414 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.296122 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.297147 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.297702 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.298818 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.299348 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.299921 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.300173 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.300945 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.301561 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.302770 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.303210 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.303785 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.304890 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.305396 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.306445 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.306886 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.308490 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.309089 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.309767 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.312082 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.312213 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.313649 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.314992 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.315830 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.316361 4944 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.316973 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.319830 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.320470 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.321482 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.323353 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.324035 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.325075 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.325786 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.327270 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.327924 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.328791 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.329108 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.329932 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.331218 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.331759 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.332982 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.333877 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.337308 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.337960 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.342128 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.352454 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.352661 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.352693 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-ovn-kubernetes\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.352749 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovn-node-metrics-cert\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.352786 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-node-log\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.352814 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-bin\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.352841 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c5899104-758a-40da-9612-7334e6616928-cni-binary-copy\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.352867 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-ovn-kubernetes\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.352867 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c5899104-758a-40da-9612-7334e6616928-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.352998 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-slash\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353031 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-kubelet\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353099 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-os-release\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353130 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-var-lib-openvswitch\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353148 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-etc-openvswitch\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353171 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-config\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353186 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-systemd-units\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353205 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-netns\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353222 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-netd\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353237 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353257 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p26h\" (UniqueName: \"kubernetes.io/projected/5ff38b2b-032d-47df-8836-105e0cfae835-kube-api-access-9p26h\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353273 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-openvswitch\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353289 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353309 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-cnibin\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353329 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-env-overrides\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353350 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-etc-kubernetes\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353366 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-ovn\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353385 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-log-socket\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353405 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-script-lib\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353423 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn57m\" (UniqueName: \"kubernetes.io/projected/aeb94161-d41a-420b-b5ed-9b65abbe91a3-kube-api-access-bn57m\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353442 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-systemd\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353459 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353475 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzdlj\" (UniqueName: \"kubernetes.io/projected/c5899104-758a-40da-9612-7334e6616928-kube-api-access-bzdlj\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353494 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-run-multus-certs\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353510 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-system-cni-dir\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353607 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-system-cni-dir\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353731 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c5899104-758a-40da-9612-7334e6616928-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353776 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353802 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-slash\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353830 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-kubelet\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.353897 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-cnibin\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354063 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-os-release\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354076 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354421 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-openvswitch\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354115 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-netd\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354151 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-netns\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354175 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-etc-openvswitch\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354543 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-systemd-units\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354767 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-etc-kubernetes\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354781 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-env-overrides\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354805 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-ovn\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354837 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-log-socket\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354870 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-bin\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354909 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-node-log\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354922 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-config\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.355281 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-systemd\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.355326 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5ff38b2b-032d-47df-8836-105e0cfae835-host-run-multus-certs\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.355633 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c5899104-758a-40da-9612-7334e6616928-tuning-conf-dir\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.354089 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-var-lib-openvswitch\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.355957 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-script-lib\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.355979 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.356613 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.357598 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovn-node-metrics-cert\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.358020 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c5899104-758a-40da-9612-7334e6616928-cni-binary-copy\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.374247 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.377145 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn57m\" (UniqueName: \"kubernetes.io/projected/aeb94161-d41a-420b-b5ed-9b65abbe91a3-kube-api-access-bn57m\") pod \"ovnkube-node-7q5tz\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.377667 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzdlj\" (UniqueName: \"kubernetes.io/projected/c5899104-758a-40da-9612-7334e6616928-kube-api-access-bzdlj\") pod \"multus-additional-cni-plugins-tc6n6\" (UID: \"c5899104-758a-40da-9612-7334e6616928\") " pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.378136 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p26h\" (UniqueName: \"kubernetes.io/projected/5ff38b2b-032d-47df-8836-105e0cfae835-kube-api-access-9p26h\") pod \"multus-dg4dk\" (UID: \"5ff38b2b-032d-47df-8836-105e0cfae835\") " pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.388003 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.391500 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.395773 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.398503 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5"} Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.399255 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.400414 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.400905 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9zglk" event={"ID":"3c792a6e-0110-4e35-923e-c1a38f32e8f9","Type":"ContainerStarted","Data":"de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073"} Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.400935 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9zglk" event={"ID":"3c792a6e-0110-4e35-923e-c1a38f32e8f9","Type":"ContainerStarted","Data":"d1b7a395d4e71cfcbb428be465f5697bfa9282e1177c6746a000ec55ed2eb211"} Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.401677 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.402746 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"22b9eb5bb8d2113045fb05aa373724efa7e76d59694115c1d52b64a8c7e6c032"} Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.409186 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.412198 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75"} Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.412252 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e"} Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.412267 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"04342ddb0833f8a21e7602530ac6e48a02bc16db8a6ed3fb9baf2d6ec4197aa3"} Nov 24 08:52:40 crc kubenswrapper[4944]: W1124 08:52:40.412498 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaeb94161_d41a_420b_b5ed_9b65abbe91a3.slice/crio-28466347089e5136be573f05db427e14a4068e3ceb329c7f08f3dc78984701a7 WatchSource:0}: Error finding container 28466347089e5136be573f05db427e14a4068e3ceb329c7f08f3dc78984701a7: Status 404 returned error can't find the container with id 28466347089e5136be573f05db427e14a4068e3ceb329c7f08f3dc78984701a7 Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.415247 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81"} Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.415292 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2d78cdb82b4603ad4f5124d780b9031a527ba0a480a14a9d1c2ce99c90da9f0c"} Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.423989 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: W1124 08:52:40.434962 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5899104_758a_40da_9612_7334e6616928.slice/crio-fdf019b12f4f10e5a862d3ec1f187ff00e6e6c5fb303eae6a61b46862e6eeae3 WatchSource:0}: Error finding container fdf019b12f4f10e5a862d3ec1f187ff00e6e6c5fb303eae6a61b46862e6eeae3: Status 404 returned error can't find the container with id fdf019b12f4f10e5a862d3ec1f187ff00e6e6c5fb303eae6a61b46862e6eeae3 Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.440433 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.459249 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.472865 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.493884 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.512306 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.529001 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.542858 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.560576 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.577529 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.595308 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.621292 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.635423 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.648383 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.670626 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.679108 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dg4dk" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.688565 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: W1124 08:52:40.693311 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ff38b2b_032d_47df_8836_105e0cfae835.slice/crio-e809109b236b9a4f8c131a8c45669fcb6f0f86508ac60226f2e77dbf55cc961e WatchSource:0}: Error finding container e809109b236b9a4f8c131a8c45669fcb6f0f86508ac60226f2e77dbf55cc961e: Status 404 returned error can't find the container with id e809109b236b9a4f8c131a8c45669fcb6f0f86508ac60226f2e77dbf55cc961e Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.713000 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.730663 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.746372 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.766980 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.782963 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.804289 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.818765 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.836164 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.849428 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.860148 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.860355 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:52:42.860315967 +0000 UTC m=+23.394756429 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.860558 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.860734 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.860798 4944 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.860926 4944 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.860955 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:42.860930066 +0000 UTC m=+23.395370528 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.861019 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:42.861008459 +0000 UTC m=+23.395449171 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.863783 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.879353 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.962248 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:40 crc kubenswrapper[4944]: I1124 08:52:40.962312 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.962556 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.962621 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.962641 4944 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.962718 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:42.962696644 +0000 UTC m=+23.497137266 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.962570 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.962842 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.962862 4944 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:40 crc kubenswrapper[4944]: E1124 08:52:40.962937 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:42.962920901 +0000 UTC m=+23.497361363 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.276136 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.276227 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.276255 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:41 crc kubenswrapper[4944]: E1124 08:52:41.276335 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:52:41 crc kubenswrapper[4944]: E1124 08:52:41.276735 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:52:41 crc kubenswrapper[4944]: E1124 08:52:41.276820 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.420894 4944 generic.go:334] "Generic (PLEG): container finished" podID="c5899104-758a-40da-9612-7334e6616928" containerID="5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e" exitCode=0 Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.420984 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" event={"ID":"c5899104-758a-40da-9612-7334e6616928","Type":"ContainerDied","Data":"5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e"} Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.421478 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" event={"ID":"c5899104-758a-40da-9612-7334e6616928","Type":"ContainerStarted","Data":"fdf019b12f4f10e5a862d3ec1f187ff00e6e6c5fb303eae6a61b46862e6eeae3"} Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.423671 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4" exitCode=0 Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.423762 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4"} Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.423799 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"28466347089e5136be573f05db427e14a4068e3ceb329c7f08f3dc78984701a7"} Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.427733 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531"} Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.427808 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83"} Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.427831 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"34735ebcfd5b005060cfb6b0d467e38b251f68a7fdf4ccb7b997f2db72936cff"} Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.438177 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dg4dk" event={"ID":"5ff38b2b-032d-47df-8836-105e0cfae835","Type":"ContainerStarted","Data":"d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6"} Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.438230 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dg4dk" event={"ID":"5ff38b2b-032d-47df-8836-105e0cfae835","Type":"ContainerStarted","Data":"e809109b236b9a4f8c131a8c45669fcb6f0f86508ac60226f2e77dbf55cc961e"} Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.453732 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.489513 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.508937 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.537985 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.554161 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.570711 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.590420 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.612916 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.629108 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.645604 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.685097 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.699643 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.719815 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.736274 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.751471 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.772313 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.788492 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.805706 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.823874 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.841302 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.859348 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.875760 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.895336 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.914663 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.938203 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.948494 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2267c"] Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.948998 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2267c" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.953292 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.953582 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.953670 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.953698 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.961725 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.974631 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9mfc\" (UniqueName: \"kubernetes.io/projected/e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b-kube-api-access-b9mfc\") pod \"node-ca-2267c\" (UID: \"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\") " pod="openshift-image-registry/node-ca-2267c" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.974715 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b-host\") pod \"node-ca-2267c\" (UID: \"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\") " pod="openshift-image-registry/node-ca-2267c" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.974778 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b-serviceca\") pod \"node-ca-2267c\" (UID: \"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\") " pod="openshift-image-registry/node-ca-2267c" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.976193 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:41 crc kubenswrapper[4944]: I1124 08:52:41.993229 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.012460 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.032882 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.065289 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.075462 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b-serviceca\") pod \"node-ca-2267c\" (UID: \"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\") " pod="openshift-image-registry/node-ca-2267c" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.075526 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9mfc\" (UniqueName: \"kubernetes.io/projected/e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b-kube-api-access-b9mfc\") pod \"node-ca-2267c\" (UID: \"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\") " pod="openshift-image-registry/node-ca-2267c" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.075576 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b-host\") pod \"node-ca-2267c\" (UID: \"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\") " pod="openshift-image-registry/node-ca-2267c" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.075669 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b-host\") pod \"node-ca-2267c\" (UID: \"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\") " pod="openshift-image-registry/node-ca-2267c" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.076655 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b-serviceca\") pod \"node-ca-2267c\" (UID: \"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\") " pod="openshift-image-registry/node-ca-2267c" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.104695 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.132838 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9mfc\" (UniqueName: \"kubernetes.io/projected/e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b-kube-api-access-b9mfc\") pod \"node-ca-2267c\" (UID: \"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\") " pod="openshift-image-registry/node-ca-2267c" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.164580 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.203438 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.242815 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.267406 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2267c" Nov 24 08:52:42 crc kubenswrapper[4944]: W1124 08:52:42.282319 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2fbc441_86ee_4abe_bfc9_75f5b8cbb04b.slice/crio-5b6608c7f1fcbdfc523c199763523bcfa827cba4d6f9467e8cb44e902c0124ae WatchSource:0}: Error finding container 5b6608c7f1fcbdfc523c199763523bcfa827cba4d6f9467e8cb44e902c0124ae: Status 404 returned error can't find the container with id 5b6608c7f1fcbdfc523c199763523bcfa827cba4d6f9467e8cb44e902c0124ae Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.284157 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.323981 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.374535 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.409239 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.443628 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.445786 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" event={"ID":"c5899104-758a-40da-9612-7334e6616928","Type":"ContainerStarted","Data":"1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e"} Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.450973 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43"} Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.451025 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152"} Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.452413 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2267c" event={"ID":"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b","Type":"ContainerStarted","Data":"5b6608c7f1fcbdfc523c199763523bcfa827cba4d6f9467e8cb44e902c0124ae"} Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.488347 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.526346 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.568368 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.608237 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.645095 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.689233 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.724394 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.766761 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.802245 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.848717 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.883080 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.883214 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.883375 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:52:46.883340179 +0000 UTC m=+27.417780641 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.883428 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.883546 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.883654 4944 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.883704 4944 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.883726 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:46.883706711 +0000 UTC m=+27.418147173 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.883745 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:46.883737822 +0000 UTC m=+27.418178284 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.921968 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.965724 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.984902 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:42 crc kubenswrapper[4944]: I1124 08:52:42.984976 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.985202 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.985233 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.985248 4944 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.985305 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:46.985287452 +0000 UTC m=+27.519727914 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.985376 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.985390 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.985400 4944 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:42 crc kubenswrapper[4944]: E1124 08:52:42.985429 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:46.985421706 +0000 UTC m=+27.519862168 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.008628 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.276830 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.276887 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:43 crc kubenswrapper[4944]: E1124 08:52:43.276978 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.276856 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:43 crc kubenswrapper[4944]: E1124 08:52:43.277175 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:52:43 crc kubenswrapper[4944]: E1124 08:52:43.277454 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.458464 4944 generic.go:334] "Generic (PLEG): container finished" podID="c5899104-758a-40da-9612-7334e6616928" containerID="1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e" exitCode=0 Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.458547 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" event={"ID":"c5899104-758a-40da-9612-7334e6616928","Type":"ContainerDied","Data":"1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e"} Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.465793 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98"} Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.465845 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597"} Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.465863 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c"} Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.465876 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f"} Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.468396 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2267c" event={"ID":"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b","Type":"ContainerStarted","Data":"f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23"} Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.470068 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d"} Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.478976 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.494708 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.510334 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.527406 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.545163 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.559717 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.574602 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.588459 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.606400 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.623288 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.644956 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.658942 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.673377 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.686464 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.705935 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.724752 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.740870 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.755165 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.771115 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.805927 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.845388 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.881387 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.931368 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:43 crc kubenswrapper[4944]: I1124 08:52:43.961824 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.005768 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.046969 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.082991 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.121601 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.475543 4944 generic.go:334] "Generic (PLEG): container finished" podID="c5899104-758a-40da-9612-7334e6616928" containerID="8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e" exitCode=0 Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.475620 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" event={"ID":"c5899104-758a-40da-9612-7334e6616928","Type":"ContainerDied","Data":"8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e"} Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.491846 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.508631 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.531496 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.545870 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.559439 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.573844 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.589865 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.606787 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.621932 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.637767 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.654222 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.669292 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.685104 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:44 crc kubenswrapper[4944]: I1124 08:52:44.699307 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:44Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.276455 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:45 crc kubenswrapper[4944]: E1124 08:52:45.276626 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.276665 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.276709 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:45 crc kubenswrapper[4944]: E1124 08:52:45.276748 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:52:45 crc kubenswrapper[4944]: E1124 08:52:45.276863 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.400331 4944 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.402921 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.402994 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.403009 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.403183 4944 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.412037 4944 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.412419 4944 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.413543 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.413587 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.413603 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.413621 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.413631 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:45Z","lastTransitionTime":"2025-11-24T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:45 crc kubenswrapper[4944]: E1124 08:52:45.427806 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.432261 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.432304 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.432314 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.432333 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.432345 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:45Z","lastTransitionTime":"2025-11-24T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:45 crc kubenswrapper[4944]: E1124 08:52:45.447145 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.451970 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.452023 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.452034 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.452071 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.452094 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:45Z","lastTransitionTime":"2025-11-24T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:45 crc kubenswrapper[4944]: E1124 08:52:45.464210 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.468639 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.468743 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.468768 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.468791 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.468807 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:45Z","lastTransitionTime":"2025-11-24T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.484560 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1"} Nov 24 08:52:45 crc kubenswrapper[4944]: E1124 08:52:45.486120 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.487631 4944 generic.go:334] "Generic (PLEG): container finished" podID="c5899104-758a-40da-9612-7334e6616928" containerID="a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4" exitCode=0 Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.487695 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" event={"ID":"c5899104-758a-40da-9612-7334e6616928","Type":"ContainerDied","Data":"a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4"} Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.490104 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.490145 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.490161 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.490178 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.490191 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:45Z","lastTransitionTime":"2025-11-24T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.505249 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: E1124 08:52:45.505496 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: E1124 08:52:45.505679 4944 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.507532 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.507570 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.507583 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.507601 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.507613 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:45Z","lastTransitionTime":"2025-11-24T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.521002 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.534809 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.549131 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.564733 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.586755 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.600765 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.613738 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.613803 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.613813 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.613833 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.613850 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:45Z","lastTransitionTime":"2025-11-24T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.617062 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.631921 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.647304 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.662930 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.678790 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.695980 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.713758 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:45Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.716794 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.716859 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.716873 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.716898 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.716916 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:45Z","lastTransitionTime":"2025-11-24T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.820145 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.820200 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.820211 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.820231 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.820247 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:45Z","lastTransitionTime":"2025-11-24T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.922953 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.923034 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.923069 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.923088 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:45 crc kubenswrapper[4944]: I1124 08:52:45.923100 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:45Z","lastTransitionTime":"2025-11-24T08:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.025884 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.025929 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.025939 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.025954 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.025965 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:46Z","lastTransitionTime":"2025-11-24T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.128774 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.128841 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.128854 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.128876 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.128892 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:46Z","lastTransitionTime":"2025-11-24T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.231946 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.232003 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.232023 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.232082 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.232101 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:46Z","lastTransitionTime":"2025-11-24T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.334479 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.334534 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.334545 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.334563 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.334575 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:46Z","lastTransitionTime":"2025-11-24T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.436973 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.437070 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.437086 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.437107 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.437121 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:46Z","lastTransitionTime":"2025-11-24T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.499241 4944 generic.go:334] "Generic (PLEG): container finished" podID="c5899104-758a-40da-9612-7334e6616928" containerID="a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a" exitCode=0 Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.499310 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" event={"ID":"c5899104-758a-40da-9612-7334e6616928","Type":"ContainerDied","Data":"a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a"} Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.519006 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.535612 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.539764 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.539859 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.539870 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.539891 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.539903 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:46Z","lastTransitionTime":"2025-11-24T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.552661 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.570142 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.585573 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.598484 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.611703 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.626193 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.639369 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.643680 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.643719 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.643732 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.643752 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.643763 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:46Z","lastTransitionTime":"2025-11-24T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.655424 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.676511 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.689611 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.704822 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.719359 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.746715 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.746809 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.746823 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.746846 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.746862 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:46Z","lastTransitionTime":"2025-11-24T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.850099 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.850143 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.850155 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.850176 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.850186 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:46Z","lastTransitionTime":"2025-11-24T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.928152 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.928283 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.928339 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:46 crc kubenswrapper[4944]: E1124 08:52:46.928438 4944 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:46 crc kubenswrapper[4944]: E1124 08:52:46.928449 4944 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:46 crc kubenswrapper[4944]: E1124 08:52:46.928499 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:54.928480595 +0000 UTC m=+35.462921057 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:46 crc kubenswrapper[4944]: E1124 08:52:46.928514 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:54.928507946 +0000 UTC m=+35.462948408 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:46 crc kubenswrapper[4944]: E1124 08:52:46.928646 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:52:54.928520366 +0000 UTC m=+35.462960868 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.953323 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.953381 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.953391 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.953408 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:46 crc kubenswrapper[4944]: I1124 08:52:46.953419 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:46Z","lastTransitionTime":"2025-11-24T08:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.029645 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.029706 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:47 crc kubenswrapper[4944]: E1124 08:52:47.029864 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:47 crc kubenswrapper[4944]: E1124 08:52:47.029888 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:47 crc kubenswrapper[4944]: E1124 08:52:47.029902 4944 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:47 crc kubenswrapper[4944]: E1124 08:52:47.029867 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:47 crc kubenswrapper[4944]: E1124 08:52:47.029964 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:55.029946593 +0000 UTC m=+35.564387055 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:47 crc kubenswrapper[4944]: E1124 08:52:47.029971 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:47 crc kubenswrapper[4944]: E1124 08:52:47.029984 4944 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:47 crc kubenswrapper[4944]: E1124 08:52:47.030022 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:55.030010465 +0000 UTC m=+35.564450927 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.056233 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.056286 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.056298 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.056318 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.056333 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:47Z","lastTransitionTime":"2025-11-24T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.161388 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.161453 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.161466 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.161489 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.161507 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:47Z","lastTransitionTime":"2025-11-24T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.265765 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.265830 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.265843 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.265865 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.265878 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:47Z","lastTransitionTime":"2025-11-24T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.276447 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.276545 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.276488 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:47 crc kubenswrapper[4944]: E1124 08:52:47.276663 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:52:47 crc kubenswrapper[4944]: E1124 08:52:47.276749 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:52:47 crc kubenswrapper[4944]: E1124 08:52:47.276840 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.368587 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.368654 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.368668 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.368692 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.368705 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:47Z","lastTransitionTime":"2025-11-24T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.471856 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.471898 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.471909 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.471935 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.471955 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:47Z","lastTransitionTime":"2025-11-24T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.511144 4944 generic.go:334] "Generic (PLEG): container finished" podID="c5899104-758a-40da-9612-7334e6616928" containerID="c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2" exitCode=0 Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.511208 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" event={"ID":"c5899104-758a-40da-9612-7334e6616928","Type":"ContainerDied","Data":"c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.518284 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.518640 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.527750 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.541748 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.555428 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.555446 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.568733 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.574370 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.574512 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.574594 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.574735 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.574861 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:47Z","lastTransitionTime":"2025-11-24T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.584420 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.604271 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.649186 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.681116 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.681610 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.681660 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.681672 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.681692 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.681705 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:47Z","lastTransitionTime":"2025-11-24T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.707041 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.725621 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.742202 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.763410 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.781710 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.784498 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.784534 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.784545 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.784567 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.784579 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:47Z","lastTransitionTime":"2025-11-24T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.798537 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.815173 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.829254 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.845947 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.863475 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.877191 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.887213 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.887277 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.887292 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.887315 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.887329 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:47Z","lastTransitionTime":"2025-11-24T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.895508 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.913603 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.929765 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.944643 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.957623 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.975496 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.990337 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.990730 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.990804 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.990896 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.990969 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:47Z","lastTransitionTime":"2025-11-24T08:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:47 crc kubenswrapper[4944]: I1124 08:52:47.991223 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.010179 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.024770 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.094725 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.095830 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.095948 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.096038 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.096232 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:48Z","lastTransitionTime":"2025-11-24T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.199324 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.199356 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.199364 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.199380 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.199390 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:48Z","lastTransitionTime":"2025-11-24T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.302217 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.302265 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.302279 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.302300 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.302313 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:48Z","lastTransitionTime":"2025-11-24T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.405996 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.406097 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.406111 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.406132 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.406148 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:48Z","lastTransitionTime":"2025-11-24T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.508575 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.508616 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.508628 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.508647 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.508661 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:48Z","lastTransitionTime":"2025-11-24T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.529232 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" event={"ID":"c5899104-758a-40da-9612-7334e6616928","Type":"ContainerStarted","Data":"5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6"} Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.529369 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.530478 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.548361 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.557114 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.569799 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.587175 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.602546 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.611679 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.611755 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.611766 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.611792 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.611804 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:48Z","lastTransitionTime":"2025-11-24T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.618494 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.641223 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.656480 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.668986 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.682155 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.696566 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.711013 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.715040 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.715120 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.715136 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.715163 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.715180 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:48Z","lastTransitionTime":"2025-11-24T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.726912 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.745099 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.761343 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.775541 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.787508 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.803884 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.817980 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.818028 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.818060 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.818080 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.818091 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:48Z","lastTransitionTime":"2025-11-24T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.820324 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.835954 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.855483 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.871534 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.888511 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.904170 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.917578 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.920680 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.920768 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.920784 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.920807 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.920823 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:48Z","lastTransitionTime":"2025-11-24T08:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.935190 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.951749 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.975401 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:48 crc kubenswrapper[4944]: I1124 08:52:48.987923 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.023479 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.023546 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.023563 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.023588 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.023605 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:49Z","lastTransitionTime":"2025-11-24T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.126619 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.126663 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.126674 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.126693 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.126705 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:49Z","lastTransitionTime":"2025-11-24T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.229897 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.229936 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.229949 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.229968 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.229982 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:49Z","lastTransitionTime":"2025-11-24T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.275993 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.276117 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:49 crc kubenswrapper[4944]: E1124 08:52:49.276203 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.276145 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:49 crc kubenswrapper[4944]: E1124 08:52:49.276336 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:52:49 crc kubenswrapper[4944]: E1124 08:52:49.276499 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.333183 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.333242 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.333259 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.333287 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.333303 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:49Z","lastTransitionTime":"2025-11-24T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.436599 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.436654 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.436664 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.436680 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.436694 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:49Z","lastTransitionTime":"2025-11-24T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.532395 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.541655 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.541712 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.541729 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.541753 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.541783 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:49Z","lastTransitionTime":"2025-11-24T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.644781 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.644856 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.644868 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.644891 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.644904 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:49Z","lastTransitionTime":"2025-11-24T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.747292 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.747355 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.747368 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.747389 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.747403 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:49Z","lastTransitionTime":"2025-11-24T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.850449 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.850496 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.850506 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.850527 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.850537 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:49Z","lastTransitionTime":"2025-11-24T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.954356 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.954438 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.954449 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.954470 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:49 crc kubenswrapper[4944]: I1124 08:52:49.954483 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:49Z","lastTransitionTime":"2025-11-24T08:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.057846 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.057906 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.057918 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.057946 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.057961 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:50Z","lastTransitionTime":"2025-11-24T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.160842 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.160903 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.160918 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.160939 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.160954 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:50Z","lastTransitionTime":"2025-11-24T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.263471 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.263516 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.263528 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.263547 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.263561 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:50Z","lastTransitionTime":"2025-11-24T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.293249 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.308955 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.322361 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.337467 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.354187 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.365673 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.365720 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.365736 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.365764 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.365778 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:50Z","lastTransitionTime":"2025-11-24T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.375617 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.388368 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.403532 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.419254 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.434955 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.453844 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.468637 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.468692 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.468705 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.468725 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.468738 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:50Z","lastTransitionTime":"2025-11-24T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.471943 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.488693 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.507532 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.536075 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.570908 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.570955 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.570964 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.570981 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.570993 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:50Z","lastTransitionTime":"2025-11-24T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.673892 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.673981 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.673994 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.674011 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.674024 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:50Z","lastTransitionTime":"2025-11-24T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.776213 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.776277 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.776290 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.776307 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.776319 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:50Z","lastTransitionTime":"2025-11-24T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.878975 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.879021 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.879031 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.879065 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.879082 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:50Z","lastTransitionTime":"2025-11-24T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.981501 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.981597 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.981620 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.981655 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:50 crc kubenswrapper[4944]: I1124 08:52:50.981708 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:50Z","lastTransitionTime":"2025-11-24T08:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.084838 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.084893 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.084904 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.084926 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.084940 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:51Z","lastTransitionTime":"2025-11-24T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.188065 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.188129 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.188144 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.188165 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.188175 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:51Z","lastTransitionTime":"2025-11-24T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.276437 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.276515 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.276633 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:51 crc kubenswrapper[4944]: E1124 08:52:51.276711 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:52:51 crc kubenswrapper[4944]: E1124 08:52:51.276900 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:52:51 crc kubenswrapper[4944]: E1124 08:52:51.277207 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.290728 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.290782 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.290795 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.290817 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.290832 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:51Z","lastTransitionTime":"2025-11-24T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.394114 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.394171 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.394189 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.394214 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.394230 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:51Z","lastTransitionTime":"2025-11-24T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.497518 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.497581 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.497592 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.497609 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.497622 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:51Z","lastTransitionTime":"2025-11-24T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.540685 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/0.log" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.544129 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6" exitCode=1 Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.544232 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6"} Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.545304 4944 scope.go:117] "RemoveContainer" containerID="6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.561301 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.576650 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.589833 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.601011 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.601117 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.601133 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.601155 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.601169 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:51Z","lastTransitionTime":"2025-11-24T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.608576 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.624385 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.639175 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.651872 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.664965 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.683031 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.695826 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.704637 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.704669 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.704677 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.704693 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.704722 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:51Z","lastTransitionTime":"2025-11-24T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.714563 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:50Z\\\",\\\"message\\\":\\\"Policy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:52:50.636823 6230 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:52:50.636845 6230 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1124 08:52:50.636894 6230 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 08:52:50.637063 6230 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:52:50.639256 6230 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:52:50.639274 6230 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:52:50.639302 6230 factory.go:656] Stopping watch factory\\\\nI1124 08:52:50.639319 6230 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:52:50.639341 6230 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:52:50.639353 6230 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 08:52:50.639360 6230 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.726808 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.740488 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.754818 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.807265 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.807316 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.807327 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.807347 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.807360 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:51Z","lastTransitionTime":"2025-11-24T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.910187 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.910236 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.910246 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.910269 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:51 crc kubenswrapper[4944]: I1124 08:52:51.910281 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:51Z","lastTransitionTime":"2025-11-24T08:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.013158 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.013235 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.013247 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.013268 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.013285 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:52Z","lastTransitionTime":"2025-11-24T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.116718 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.116778 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.116794 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.116818 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.116855 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:52Z","lastTransitionTime":"2025-11-24T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.221065 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.221103 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.221112 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.221127 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.221138 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:52Z","lastTransitionTime":"2025-11-24T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.331302 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.331338 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.331347 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.331365 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.331374 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:52Z","lastTransitionTime":"2025-11-24T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.434490 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.434543 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.434553 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.434569 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.434581 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:52Z","lastTransitionTime":"2025-11-24T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.537157 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.537307 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.537429 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.537452 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.537532 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:52Z","lastTransitionTime":"2025-11-24T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.549589 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/0.log" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.553253 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.553461 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.569880 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.584228 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.601701 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.616909 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.635920 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.639881 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.639950 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.639965 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.639988 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.640011 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:52Z","lastTransitionTime":"2025-11-24T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.651895 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.668562 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.682478 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.696424 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.710701 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.726087 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.735227 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d"] Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.735948 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.738400 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.738675 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.743811 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.743871 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.743885 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.743917 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.743932 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:52Z","lastTransitionTime":"2025-11-24T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.754563 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:50Z\\\",\\\"message\\\":\\\"Policy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:52:50.636823 6230 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:52:50.636845 6230 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1124 08:52:50.636894 6230 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 08:52:50.637063 6230 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:52:50.639256 6230 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:52:50.639274 6230 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:52:50.639302 6230 factory.go:656] Stopping watch factory\\\\nI1124 08:52:50.639319 6230 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:52:50.639341 6230 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:52:50.639353 6230 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 08:52:50.639360 6230 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.772877 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.786384 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.801183 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.813378 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.827665 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.834514 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8bdk\" (UniqueName: \"kubernetes.io/projected/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-kube-api-access-f8bdk\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.834574 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.834693 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.834724 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-env-overrides\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.841177 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.846247 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.846293 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.846306 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.846336 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.846352 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:52Z","lastTransitionTime":"2025-11-24T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.854600 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.867788 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.881083 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.894510 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.907946 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.920027 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.935093 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.935769 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-env-overrides\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.935816 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8bdk\" (UniqueName: \"kubernetes.io/projected/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-kube-api-access-f8bdk\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.935899 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.936005 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.936581 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-env-overrides\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.936909 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.942888 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.948800 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.948900 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.948913 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.948934 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.948949 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:52Z","lastTransitionTime":"2025-11-24T08:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.954363 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.956904 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8bdk\" (UniqueName: \"kubernetes.io/projected/fbcff1fc-7f00-49e6-8095-fde5f3c24a57-kube-api-access-f8bdk\") pod \"ovnkube-control-plane-749d76644c-llb5d\" (UID: \"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.974452 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:50Z\\\",\\\"message\\\":\\\"Policy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:52:50.636823 6230 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:52:50.636845 6230 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1124 08:52:50.636894 6230 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 08:52:50.637063 6230 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:52:50.639256 6230 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:52:50.639274 6230 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:52:50.639302 6230 factory.go:656] Stopping watch factory\\\\nI1124 08:52:50.639319 6230 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:52:50.639341 6230 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:52:50.639353 6230 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 08:52:50.639360 6230 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:52 crc kubenswrapper[4944]: I1124 08:52:52.986548 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.002928 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.051429 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.051481 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.051491 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.051510 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.051522 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:53Z","lastTransitionTime":"2025-11-24T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.054944 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.154391 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.154431 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.154442 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.154459 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.154470 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:53Z","lastTransitionTime":"2025-11-24T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.257349 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.257418 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.257472 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.257500 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.257515 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:53Z","lastTransitionTime":"2025-11-24T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.276816 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.276817 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:53 crc kubenswrapper[4944]: E1124 08:52:53.276955 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.276988 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:53 crc kubenswrapper[4944]: E1124 08:52:53.277040 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:52:53 crc kubenswrapper[4944]: E1124 08:52:53.277167 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.360602 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.360648 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.360657 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.360673 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.360686 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:53Z","lastTransitionTime":"2025-11-24T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.463883 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.463933 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.463943 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.463963 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.463976 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:53Z","lastTransitionTime":"2025-11-24T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.489007 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-jmkb9"] Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.489745 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:53 crc kubenswrapper[4944]: E1124 08:52:53.489844 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.508889 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.536509 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.550589 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.558743 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" event={"ID":"fbcff1fc-7f00-49e6-8095-fde5f3c24a57","Type":"ContainerStarted","Data":"8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.558826 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" event={"ID":"fbcff1fc-7f00-49e6-8095-fde5f3c24a57","Type":"ContainerStarted","Data":"de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.558845 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" event={"ID":"fbcff1fc-7f00-49e6-8095-fde5f3c24a57","Type":"ContainerStarted","Data":"4724e6b1cae4affe45be42eb225817ea14ffd92054079f16f561d27db1394f11"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.560881 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/1.log" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.561607 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/0.log" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.564634 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b" exitCode=1 Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.564680 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.564752 4944 scope.go:117] "RemoveContainer" containerID="6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.565542 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.565570 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.565580 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.565595 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.565606 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:53Z","lastTransitionTime":"2025-11-24T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.565735 4944 scope.go:117] "RemoveContainer" containerID="7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b" Nov 24 08:52:53 crc kubenswrapper[4944]: E1124 08:52:53.565966 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.570951 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.589418 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.604690 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.619515 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.634293 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.647489 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.647553 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsbbf\" (UniqueName: \"kubernetes.io/projected/55990e0c-a6ff-4b18-8b8d-f27542761408-kube-api-access-fsbbf\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.649965 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.664597 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.668737 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.668780 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.668790 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.668809 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.668825 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:53Z","lastTransitionTime":"2025-11-24T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.683689 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:50Z\\\",\\\"message\\\":\\\"Policy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:52:50.636823 6230 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:52:50.636845 6230 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1124 08:52:50.636894 6230 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 08:52:50.637063 6230 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:52:50.639256 6230 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:52:50.639274 6230 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:52:50.639302 6230 factory.go:656] Stopping watch factory\\\\nI1124 08:52:50.639319 6230 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:52:50.639341 6230 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:52:50.639353 6230 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 08:52:50.639360 6230 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.696423 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.722558 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.749457 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.749506 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsbbf\" (UniqueName: \"kubernetes.io/projected/55990e0c-a6ff-4b18-8b8d-f27542761408-kube-api-access-fsbbf\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:53 crc kubenswrapper[4944]: E1124 08:52:53.750002 4944 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:52:53 crc kubenswrapper[4944]: E1124 08:52:53.750122 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs podName:55990e0c-a6ff-4b18-8b8d-f27542761408 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:54.250094143 +0000 UTC m=+34.784534605 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs") pod "network-metrics-daemon-jmkb9" (UID: "55990e0c-a6ff-4b18-8b8d-f27542761408") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.752213 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.773017 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.773089 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.773102 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.773121 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.773442 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:53Z","lastTransitionTime":"2025-11-24T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.773833 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.774340 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsbbf\" (UniqueName: \"kubernetes.io/projected/55990e0c-a6ff-4b18-8b8d-f27542761408-kube-api-access-fsbbf\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.788011 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.800729 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.815922 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.834660 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.857234 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:50Z\\\",\\\"message\\\":\\\"Policy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:52:50.636823 6230 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:52:50.636845 6230 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1124 08:52:50.636894 6230 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 08:52:50.637063 6230 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:52:50.639256 6230 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:52:50.639274 6230 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:52:50.639302 6230 factory.go:656] Stopping watch factory\\\\nI1124 08:52:50.639319 6230 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:52:50.639341 6230 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:52:50.639353 6230 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 08:52:50.639360 6230 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.871297 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.875889 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.875945 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.875960 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.875982 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.875995 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:53Z","lastTransitionTime":"2025-11-24T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.885540 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.900225 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.913939 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.929993 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.942942 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.956609 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.971483 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.979107 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.979158 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.979169 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.979235 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.979252 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:53Z","lastTransitionTime":"2025-11-24T08:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:53 crc kubenswrapper[4944]: I1124 08:52:53.987525 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.001377 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.014824 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.029067 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.081860 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.081904 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.081919 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.081938 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.081950 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:54Z","lastTransitionTime":"2025-11-24T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.185089 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.185137 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.185147 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.185167 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.185180 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:54Z","lastTransitionTime":"2025-11-24T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.254654 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:54 crc kubenswrapper[4944]: E1124 08:52:54.254883 4944 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:52:54 crc kubenswrapper[4944]: E1124 08:52:54.254967 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs podName:55990e0c-a6ff-4b18-8b8d-f27542761408 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:55.254947215 +0000 UTC m=+35.789387667 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs") pod "network-metrics-daemon-jmkb9" (UID: "55990e0c-a6ff-4b18-8b8d-f27542761408") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.287870 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.287911 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.287922 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.287938 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.287950 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:54Z","lastTransitionTime":"2025-11-24T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.390845 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.390890 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.390902 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.390919 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.390928 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:54Z","lastTransitionTime":"2025-11-24T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.494497 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.494540 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.494549 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.494565 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.494578 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:54Z","lastTransitionTime":"2025-11-24T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.570676 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/1.log" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.598108 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.598153 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.598165 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.598183 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.598194 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:54Z","lastTransitionTime":"2025-11-24T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.700391 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.700436 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.700450 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.700476 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.700496 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:54Z","lastTransitionTime":"2025-11-24T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.785216 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.803724 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.803778 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.803788 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.803807 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.803822 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:54Z","lastTransitionTime":"2025-11-24T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.804851 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.820602 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.843937 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6cf424ab998b80264c17187d1ad4b0f2325dc2e09aaa9251e0713a3af63219c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:50Z\\\",\\\"message\\\":\\\"Policy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:52:50.636823 6230 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:52:50.636845 6230 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1124 08:52:50.636894 6230 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 08:52:50.637063 6230 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:52:50.639256 6230 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:52:50.639274 6230 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:52:50.639302 6230 factory.go:656] Stopping watch factory\\\\nI1124 08:52:50.639319 6230 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:52:50.639341 6230 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:52:50.639353 6230 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 08:52:50.639360 6230 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.858268 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.874082 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.890397 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.904032 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.906494 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.906541 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.906553 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.906573 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.906584 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:54Z","lastTransitionTime":"2025-11-24T08:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.918111 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.935341 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.951069 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.963088 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.963394 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.963437 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:54 crc kubenswrapper[4944]: E1124 08:52:54.963571 4944 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:54 crc kubenswrapper[4944]: E1124 08:52:54.963642 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:10.963625171 +0000 UTC m=+51.498065633 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:52:54 crc kubenswrapper[4944]: E1124 08:52:54.963788 4944 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:54 crc kubenswrapper[4944]: E1124 08:52:54.963844 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:10.963830128 +0000 UTC m=+51.498270590 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:52:54 crc kubenswrapper[4944]: E1124 08:52:54.963955 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:53:10.963922991 +0000 UTC m=+51.498363453 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.965969 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:54 crc kubenswrapper[4944]: I1124 08:52:54.986128 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.009280 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.009332 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.009346 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.009381 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.009396 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.010016 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.026066 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.041563 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.055652 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.064135 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.064241 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.064405 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.064431 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.064445 4944 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.064405 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.064539 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.064555 4944 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.064509 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:11.06449021 +0000 UTC m=+51.598930672 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.064605 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:11.064592214 +0000 UTC m=+51.599032676 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.112497 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.113382 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.113454 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.113486 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.113506 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.217405 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.217475 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.217486 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.217505 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.217517 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.266484 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.266771 4944 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.267056 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs podName:55990e0c-a6ff-4b18-8b8d-f27542761408 nodeName:}" failed. No retries permitted until 2025-11-24 08:52:57.267016053 +0000 UTC m=+37.801456515 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs") pod "network-metrics-daemon-jmkb9" (UID: "55990e0c-a6ff-4b18-8b8d-f27542761408") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.276865 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.276945 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.276901 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.276865 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.277039 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.277162 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.277292 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.277412 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.320307 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.320343 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.320352 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.320376 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.320389 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.424406 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.424457 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.424474 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.424494 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.424506 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.527594 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.527633 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.527643 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.527660 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.527671 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.631418 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.631463 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.631477 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.631495 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.631508 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.734329 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.734410 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.734421 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.734438 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.734450 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.836760 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.836793 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.836802 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.836818 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.836828 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.865189 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.865246 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.865258 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.865276 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.865286 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.880234 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.885577 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.885619 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.885627 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.885643 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.885664 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.900497 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.905956 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.906010 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.906023 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.906068 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.906089 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.920569 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.926070 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.926123 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.926141 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.926160 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.926173 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.939961 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.945637 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.945719 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.945734 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.945777 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.945792 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.962861 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:55 crc kubenswrapper[4944]: E1124 08:52:55.963027 4944 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.965680 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.965811 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.965832 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.965857 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:55 crc kubenswrapper[4944]: I1124 08:52:55.965915 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:55Z","lastTransitionTime":"2025-11-24T08:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.069598 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.069646 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.069655 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.069675 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.069686 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:56Z","lastTransitionTime":"2025-11-24T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.172279 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.172324 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.172336 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.172360 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.172374 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:56Z","lastTransitionTime":"2025-11-24T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.275775 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.275831 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.275840 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.275854 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.275866 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:56Z","lastTransitionTime":"2025-11-24T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.379236 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.379305 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.379318 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.379348 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.379364 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:56Z","lastTransitionTime":"2025-11-24T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.482226 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.482270 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.482280 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.482296 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.482327 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:56Z","lastTransitionTime":"2025-11-24T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.584724 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.584793 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.584833 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.584857 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.584868 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:56Z","lastTransitionTime":"2025-11-24T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.687957 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.688007 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.688021 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.688061 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.688077 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:56Z","lastTransitionTime":"2025-11-24T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.800365 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.800421 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.800470 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.800496 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.800531 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:56Z","lastTransitionTime":"2025-11-24T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.903312 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.903376 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.903389 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.903407 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:56 crc kubenswrapper[4944]: I1124 08:52:56.903419 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:56Z","lastTransitionTime":"2025-11-24T08:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.006728 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.006792 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.006806 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.006831 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.006848 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:57Z","lastTransitionTime":"2025-11-24T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.109257 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.109306 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.109315 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.109332 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.109343 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:57Z","lastTransitionTime":"2025-11-24T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.212212 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.212261 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.212271 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.212289 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.212301 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:57Z","lastTransitionTime":"2025-11-24T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.276234 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.276303 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.276376 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.276340 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:57 crc kubenswrapper[4944]: E1124 08:52:57.276456 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:52:57 crc kubenswrapper[4944]: E1124 08:52:57.276573 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:52:57 crc kubenswrapper[4944]: E1124 08:52:57.276674 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:52:57 crc kubenswrapper[4944]: E1124 08:52:57.276717 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.294227 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:57 crc kubenswrapper[4944]: E1124 08:52:57.294404 4944 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:52:57 crc kubenswrapper[4944]: E1124 08:52:57.294459 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs podName:55990e0c-a6ff-4b18-8b8d-f27542761408 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:01.294440335 +0000 UTC m=+41.828880797 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs") pod "network-metrics-daemon-jmkb9" (UID: "55990e0c-a6ff-4b18-8b8d-f27542761408") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.315862 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.315928 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.315940 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.315960 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.315976 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:57Z","lastTransitionTime":"2025-11-24T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.419432 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.419492 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.419514 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.419545 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.419566 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:57Z","lastTransitionTime":"2025-11-24T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.522356 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.522418 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.522427 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.522444 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.522454 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:57Z","lastTransitionTime":"2025-11-24T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.624996 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.625115 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.625125 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.625146 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.625160 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:57Z","lastTransitionTime":"2025-11-24T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.728351 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.728414 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.728427 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.728447 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.728460 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:57Z","lastTransitionTime":"2025-11-24T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.831166 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.831229 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.831240 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.831261 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.831278 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:57Z","lastTransitionTime":"2025-11-24T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.933899 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.933956 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.933967 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.933989 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:57 crc kubenswrapper[4944]: I1124 08:52:57.934002 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:57Z","lastTransitionTime":"2025-11-24T08:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.036896 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.036947 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.036956 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.036972 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.036982 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:58Z","lastTransitionTime":"2025-11-24T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.140322 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.140374 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.140385 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.140404 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.140416 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:58Z","lastTransitionTime":"2025-11-24T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.243601 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.243648 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.243660 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.243679 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.243692 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:58Z","lastTransitionTime":"2025-11-24T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.346484 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.346572 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.346587 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.346608 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.346624 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:58Z","lastTransitionTime":"2025-11-24T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.449630 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.449682 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.449701 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.449725 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.449740 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:58Z","lastTransitionTime":"2025-11-24T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.552341 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.552389 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.552399 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.552417 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.552427 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:58Z","lastTransitionTime":"2025-11-24T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.655747 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.655801 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.655812 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.655831 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.655843 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:58Z","lastTransitionTime":"2025-11-24T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.758926 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.758972 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.758982 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.759000 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.759015 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:58Z","lastTransitionTime":"2025-11-24T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.862186 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.862239 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.862251 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.862270 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.862284 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:58Z","lastTransitionTime":"2025-11-24T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.964952 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.965002 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.965010 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.965032 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:58 crc kubenswrapper[4944]: I1124 08:52:58.965055 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:58Z","lastTransitionTime":"2025-11-24T08:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.067871 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.067918 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.067931 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.067950 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.067964 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:59Z","lastTransitionTime":"2025-11-24T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.170999 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.171118 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.171134 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.171161 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.171177 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:59Z","lastTransitionTime":"2025-11-24T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.190161 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.191081 4944 scope.go:117] "RemoveContainer" containerID="7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b" Nov 24 08:52:59 crc kubenswrapper[4944]: E1124 08:52:59.191260 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.208530 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.223783 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.245470 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.258889 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.274436 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.274924 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.274960 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.274973 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.274994 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.275007 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:59Z","lastTransitionTime":"2025-11-24T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.275943 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.275990 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.275994 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:52:59 crc kubenswrapper[4944]: E1124 08:52:59.276084 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:52:59 crc kubenswrapper[4944]: E1124 08:52:59.276397 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.276409 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:52:59 crc kubenswrapper[4944]: E1124 08:52:59.276510 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:52:59 crc kubenswrapper[4944]: E1124 08:52:59.276595 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.286669 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.299994 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.315311 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.330493 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.346090 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.361879 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.379024 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.379534 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.379625 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.379064 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.379721 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.379903 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:59Z","lastTransitionTime":"2025-11-24T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.397626 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.410628 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.424181 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.438434 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:52:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.483123 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.483167 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.483179 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.483202 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.483213 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:59Z","lastTransitionTime":"2025-11-24T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.586110 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.586163 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.586176 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.586199 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.586214 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:59Z","lastTransitionTime":"2025-11-24T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.689267 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.689353 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.689367 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.689389 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.689403 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:59Z","lastTransitionTime":"2025-11-24T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.791782 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.791835 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.791846 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.791867 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.791879 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:59Z","lastTransitionTime":"2025-11-24T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.895499 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.895547 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.895559 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.895578 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.895590 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:59Z","lastTransitionTime":"2025-11-24T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.997961 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.997996 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.998005 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.998020 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:52:59 crc kubenswrapper[4944]: I1124 08:52:59.998030 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:52:59Z","lastTransitionTime":"2025-11-24T08:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.100626 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.100670 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.100680 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.100695 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.100705 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:00Z","lastTransitionTime":"2025-11-24T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.203152 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.203194 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.203208 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.203224 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.203238 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:00Z","lastTransitionTime":"2025-11-24T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.290730 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.303521 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.305608 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.305640 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.305649 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.305665 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.305676 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:00Z","lastTransitionTime":"2025-11-24T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.315672 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.327775 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.341072 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.360842 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.372093 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.384474 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.398246 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.411256 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.411322 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.411334 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.411366 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.411377 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:00Z","lastTransitionTime":"2025-11-24T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.413361 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.425971 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.441111 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.456457 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.468955 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.483264 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.500982 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.514019 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.514102 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.514115 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.514138 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.514150 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:00Z","lastTransitionTime":"2025-11-24T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.617172 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.617222 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.617234 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.617254 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.617266 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:00Z","lastTransitionTime":"2025-11-24T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.720683 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.720735 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.720747 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.720767 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.720781 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:00Z","lastTransitionTime":"2025-11-24T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.823928 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.823982 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.824021 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.824068 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.824085 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:00Z","lastTransitionTime":"2025-11-24T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.927015 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.927081 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.927092 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.927111 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:00 crc kubenswrapper[4944]: I1124 08:53:00.927120 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:00Z","lastTransitionTime":"2025-11-24T08:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.029762 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.029853 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.029867 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.029885 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.029897 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:01Z","lastTransitionTime":"2025-11-24T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.132683 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.132726 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.132736 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.132753 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.132765 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:01Z","lastTransitionTime":"2025-11-24T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.236677 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.236728 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.236741 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.236764 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.236779 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:01Z","lastTransitionTime":"2025-11-24T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.276412 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:01 crc kubenswrapper[4944]: E1124 08:53:01.276599 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.277123 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.277228 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:01 crc kubenswrapper[4944]: E1124 08:53:01.277372 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.277245 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:01 crc kubenswrapper[4944]: E1124 08:53:01.277572 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:01 crc kubenswrapper[4944]: E1124 08:53:01.277681 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.339599 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.339646 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.339657 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.339676 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.339687 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:01Z","lastTransitionTime":"2025-11-24T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.340035 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:01 crc kubenswrapper[4944]: E1124 08:53:01.340296 4944 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:53:01 crc kubenswrapper[4944]: E1124 08:53:01.340430 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs podName:55990e0c-a6ff-4b18-8b8d-f27542761408 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:09.340408167 +0000 UTC m=+49.874848629 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs") pod "network-metrics-daemon-jmkb9" (UID: "55990e0c-a6ff-4b18-8b8d-f27542761408") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.442441 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.442513 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.442525 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.442563 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.442576 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:01Z","lastTransitionTime":"2025-11-24T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.545009 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.545088 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.545104 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.545129 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.545145 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:01Z","lastTransitionTime":"2025-11-24T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.648142 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.648195 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.648206 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.648228 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.648244 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:01Z","lastTransitionTime":"2025-11-24T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.751661 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.751705 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.751715 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.751734 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.751745 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:01Z","lastTransitionTime":"2025-11-24T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.854782 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.854833 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.854845 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.854863 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.854874 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:01Z","lastTransitionTime":"2025-11-24T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.958641 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.958720 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.958743 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.958790 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:01 crc kubenswrapper[4944]: I1124 08:53:01.958815 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:01Z","lastTransitionTime":"2025-11-24T08:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.061895 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.062263 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.062354 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.062434 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.062530 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:02Z","lastTransitionTime":"2025-11-24T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.165707 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.165755 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.165765 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.165786 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.165798 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:02Z","lastTransitionTime":"2025-11-24T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.269245 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.269290 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.269301 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.269320 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.269333 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:02Z","lastTransitionTime":"2025-11-24T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.373472 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.373542 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.373567 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.373598 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.373622 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:02Z","lastTransitionTime":"2025-11-24T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.475917 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.475982 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.475993 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.476012 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.476022 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:02Z","lastTransitionTime":"2025-11-24T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.579031 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.579129 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.579148 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.579199 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.579213 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:02Z","lastTransitionTime":"2025-11-24T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.682158 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.682230 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.682242 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.682262 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.682276 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:02Z","lastTransitionTime":"2025-11-24T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.785175 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.785228 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.785239 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.785265 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.785282 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:02Z","lastTransitionTime":"2025-11-24T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.887513 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.887557 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.887567 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.887583 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.887595 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:02Z","lastTransitionTime":"2025-11-24T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.991207 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.991252 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.991267 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.991286 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:02 crc kubenswrapper[4944]: I1124 08:53:02.991299 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:02Z","lastTransitionTime":"2025-11-24T08:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.095074 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.095165 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.095182 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.095202 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.095213 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:03Z","lastTransitionTime":"2025-11-24T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.198379 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.198782 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.199248 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.199399 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.199514 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:03Z","lastTransitionTime":"2025-11-24T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.276511 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.276521 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.276537 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.276700 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:03 crc kubenswrapper[4944]: E1124 08:53:03.276932 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:03 crc kubenswrapper[4944]: E1124 08:53:03.277147 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:03 crc kubenswrapper[4944]: E1124 08:53:03.277170 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:03 crc kubenswrapper[4944]: E1124 08:53:03.277258 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.303015 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.303091 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.303105 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.303126 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.303139 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:03Z","lastTransitionTime":"2025-11-24T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.406812 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.407306 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.407548 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.407707 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.407827 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:03Z","lastTransitionTime":"2025-11-24T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.511197 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.511248 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.511257 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.511274 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.511286 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:03Z","lastTransitionTime":"2025-11-24T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.614506 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.614873 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.614989 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.615158 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.615290 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:03Z","lastTransitionTime":"2025-11-24T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.718316 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.718359 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.718371 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.718395 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.718409 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:03Z","lastTransitionTime":"2025-11-24T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.821632 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.821696 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.821708 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.821729 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.821741 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:03Z","lastTransitionTime":"2025-11-24T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.924532 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.924598 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.924619 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.924646 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:03 crc kubenswrapper[4944]: I1124 08:53:03.924660 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:03Z","lastTransitionTime":"2025-11-24T08:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.028106 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.028161 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.028171 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.028190 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.028203 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:04Z","lastTransitionTime":"2025-11-24T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.131160 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.131207 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.131217 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.131234 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.131245 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:04Z","lastTransitionTime":"2025-11-24T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.234760 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.234828 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.234854 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.234885 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.234898 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:04Z","lastTransitionTime":"2025-11-24T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.337768 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.337821 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.337833 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.337858 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.337868 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:04Z","lastTransitionTime":"2025-11-24T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.440682 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.440741 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.440754 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.440778 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.440794 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:04Z","lastTransitionTime":"2025-11-24T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.543987 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.544038 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.544074 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.544092 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.544106 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:04Z","lastTransitionTime":"2025-11-24T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.647119 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.647219 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.647235 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.647262 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.647276 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:04Z","lastTransitionTime":"2025-11-24T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.750830 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.750934 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.750948 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.750972 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.750988 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:04Z","lastTransitionTime":"2025-11-24T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.854697 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.854751 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.854761 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.854780 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.854792 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:04Z","lastTransitionTime":"2025-11-24T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.957532 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.957585 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.957603 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.957624 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:04 crc kubenswrapper[4944]: I1124 08:53:04.957640 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:04Z","lastTransitionTime":"2025-11-24T08:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.059894 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.059939 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.059951 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.059968 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.059978 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:05Z","lastTransitionTime":"2025-11-24T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.163467 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.163754 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.163827 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.163947 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.164007 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:05Z","lastTransitionTime":"2025-11-24T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.266630 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.266675 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.266686 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.266704 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.266716 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:05Z","lastTransitionTime":"2025-11-24T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.275957 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:05 crc kubenswrapper[4944]: E1124 08:53:05.276381 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.275983 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:05 crc kubenswrapper[4944]: E1124 08:53:05.276609 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.275957 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:05 crc kubenswrapper[4944]: E1124 08:53:05.276805 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.276952 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:05 crc kubenswrapper[4944]: E1124 08:53:05.277117 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.368962 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.369022 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.369034 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.369077 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.369091 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:05Z","lastTransitionTime":"2025-11-24T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.471550 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.471813 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.471894 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.471957 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.472022 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:05Z","lastTransitionTime":"2025-11-24T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.575162 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.575228 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.575251 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.575277 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.575294 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:05Z","lastTransitionTime":"2025-11-24T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.678470 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.678513 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.678522 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.678538 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.678551 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:05Z","lastTransitionTime":"2025-11-24T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.782119 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.782161 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.782172 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.782190 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.782203 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:05Z","lastTransitionTime":"2025-11-24T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.884899 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.884990 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.885008 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.885028 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.885086 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:05Z","lastTransitionTime":"2025-11-24T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.987964 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.988014 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.988025 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.988043 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:05 crc kubenswrapper[4944]: I1124 08:53:05.988078 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:05Z","lastTransitionTime":"2025-11-24T08:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.090774 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.091096 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.091252 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.091312 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.091327 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.194667 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.194727 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.194737 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.194762 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.194776 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.259671 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.259725 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.259736 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.259758 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.259781 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: E1124 08:53:06.272011 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:06Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.276353 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.276382 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.276391 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.276405 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.276417 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: E1124 08:53:06.289502 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:06Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.294527 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.294572 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.294584 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.294601 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.294613 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: E1124 08:53:06.308582 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:06Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.312938 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.312988 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.313001 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.313025 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.313062 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: E1124 08:53:06.325920 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:06Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.329760 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.329802 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.329815 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.329832 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.329844 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: E1124 08:53:06.341736 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:06Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:06 crc kubenswrapper[4944]: E1124 08:53:06.341849 4944 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.343828 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.343869 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.343879 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.343899 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.343919 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.446921 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.446953 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.446963 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.446978 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.446989 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.552623 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.552665 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.552681 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.552697 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.552708 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.655262 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.655533 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.655633 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.655728 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.655831 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.759578 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.759630 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.759640 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.759657 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.759673 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.862374 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.862434 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.862445 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.862464 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.862476 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.964621 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.964663 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.964672 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.964686 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:06 crc kubenswrapper[4944]: I1124 08:53:06.964696 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:06Z","lastTransitionTime":"2025-11-24T08:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.067488 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.067533 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.067545 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.067561 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.067573 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:07Z","lastTransitionTime":"2025-11-24T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.170104 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.170155 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.170165 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.170190 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.170202 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:07Z","lastTransitionTime":"2025-11-24T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.274350 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.274421 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.274436 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.274459 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.274476 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:07Z","lastTransitionTime":"2025-11-24T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.276662 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.276673 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.276781 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.276863 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:07 crc kubenswrapper[4944]: E1124 08:53:07.277022 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:07 crc kubenswrapper[4944]: E1124 08:53:07.277121 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:07 crc kubenswrapper[4944]: E1124 08:53:07.277242 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:07 crc kubenswrapper[4944]: E1124 08:53:07.277486 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.377628 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.377711 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.377729 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.377753 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.377770 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:07Z","lastTransitionTime":"2025-11-24T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.480785 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.480832 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.480842 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.480858 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.480885 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:07Z","lastTransitionTime":"2025-11-24T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.584023 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.584477 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.584553 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.584632 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.584721 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:07Z","lastTransitionTime":"2025-11-24T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.687532 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.687606 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.687617 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.687640 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.687660 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:07Z","lastTransitionTime":"2025-11-24T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.790140 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.790204 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.790219 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.790240 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.790256 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:07Z","lastTransitionTime":"2025-11-24T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.892855 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.892907 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.892917 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.892936 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.892949 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:07Z","lastTransitionTime":"2025-11-24T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.996614 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.996727 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.996748 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.996769 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:07 crc kubenswrapper[4944]: I1124 08:53:07.996780 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:07Z","lastTransitionTime":"2025-11-24T08:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.099723 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.099767 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.099775 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.099791 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.099801 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:08Z","lastTransitionTime":"2025-11-24T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.203536 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.203596 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.203615 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.203646 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.203662 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:08Z","lastTransitionTime":"2025-11-24T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.306149 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.306500 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.306579 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.306645 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.306708 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:08Z","lastTransitionTime":"2025-11-24T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.409644 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.409933 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.410026 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.410148 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.410211 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:08Z","lastTransitionTime":"2025-11-24T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.512331 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.512681 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.512757 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.512833 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.512936 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:08Z","lastTransitionTime":"2025-11-24T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.616483 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.616940 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.617013 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.617103 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.617177 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:08Z","lastTransitionTime":"2025-11-24T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.720160 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.720209 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.720221 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.720241 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.720254 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:08Z","lastTransitionTime":"2025-11-24T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.823274 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.823353 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.823365 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.823389 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.823405 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:08Z","lastTransitionTime":"2025-11-24T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.930541 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.930881 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.930959 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.931035 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:08 crc kubenswrapper[4944]: I1124 08:53:08.931134 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:08Z","lastTransitionTime":"2025-11-24T08:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.035965 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.036023 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.036035 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.036079 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.036094 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:09Z","lastTransitionTime":"2025-11-24T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.138917 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.138970 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.138983 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.139006 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.139023 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:09Z","lastTransitionTime":"2025-11-24T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.242464 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.242525 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.242536 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.242558 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.242568 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:09Z","lastTransitionTime":"2025-11-24T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.275922 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.276370 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.276119 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.276495 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:09 crc kubenswrapper[4944]: E1124 08:53:09.276694 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:09 crc kubenswrapper[4944]: E1124 08:53:09.276795 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:09 crc kubenswrapper[4944]: E1124 08:53:09.276850 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:09 crc kubenswrapper[4944]: E1124 08:53:09.276918 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.345430 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.345503 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.345516 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.345536 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.345566 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:09Z","lastTransitionTime":"2025-11-24T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.440109 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:09 crc kubenswrapper[4944]: E1124 08:53:09.440316 4944 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:53:09 crc kubenswrapper[4944]: E1124 08:53:09.440384 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs podName:55990e0c-a6ff-4b18-8b8d-f27542761408 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:25.440366505 +0000 UTC m=+65.974806967 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs") pod "network-metrics-daemon-jmkb9" (UID: "55990e0c-a6ff-4b18-8b8d-f27542761408") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.448787 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.449102 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.449351 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.449444 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.449681 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:09Z","lastTransitionTime":"2025-11-24T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.552616 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.552656 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.552665 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.552680 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.552693 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:09Z","lastTransitionTime":"2025-11-24T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.656082 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.656168 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.656183 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.656201 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.656212 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:09Z","lastTransitionTime":"2025-11-24T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.759434 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.759480 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.759498 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.759520 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.759538 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:09Z","lastTransitionTime":"2025-11-24T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.862558 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.862606 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.862617 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.862634 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.862646 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:09Z","lastTransitionTime":"2025-11-24T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.965504 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.965568 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.965585 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.965606 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:09 crc kubenswrapper[4944]: I1124 08:53:09.965620 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:09Z","lastTransitionTime":"2025-11-24T08:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.068732 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.068800 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.068814 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.068836 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.068851 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:10Z","lastTransitionTime":"2025-11-24T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.171952 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.171988 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.171997 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.172013 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.172024 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:10Z","lastTransitionTime":"2025-11-24T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.274949 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.274992 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.275006 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.275025 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.275040 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:10Z","lastTransitionTime":"2025-11-24T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.292987 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.309174 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.322785 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.334770 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.349296 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.372009 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.378770 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.378849 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.378864 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.378888 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.378902 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:10Z","lastTransitionTime":"2025-11-24T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.394277 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.408381 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.422462 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.435305 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.449363 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.464204 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.478442 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.487709 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.487756 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.487767 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.487784 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.487865 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:10Z","lastTransitionTime":"2025-11-24T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.516905 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.542753 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.564671 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:10Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.591515 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.591568 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.591583 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.591605 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.591620 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:10Z","lastTransitionTime":"2025-11-24T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.695163 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.695586 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.695701 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.695805 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.695895 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:10Z","lastTransitionTime":"2025-11-24T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.798706 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.798756 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.798766 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.798785 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.798798 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:10Z","lastTransitionTime":"2025-11-24T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.901886 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.901943 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.901955 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.901973 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:10 crc kubenswrapper[4944]: I1124 08:53:10.901986 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:10Z","lastTransitionTime":"2025-11-24T08:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.004958 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.005028 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.005089 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.005127 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.005150 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:11Z","lastTransitionTime":"2025-11-24T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.055634 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.055759 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.055807 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.055905 4944 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.055907 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:53:43.055874492 +0000 UTC m=+83.590314994 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.055977 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:43.055964185 +0000 UTC m=+83.590404677 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.056775 4944 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.056839 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:43.056826552 +0000 UTC m=+83.591267014 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.108586 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.108636 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.108652 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.108674 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.108685 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:11Z","lastTransitionTime":"2025-11-24T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.157328 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.157389 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.157563 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.157581 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.157594 4944 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.157661 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:43.157644659 +0000 UTC m=+83.692085121 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.157725 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.157788 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.157811 4944 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.157914 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:43.157879517 +0000 UTC m=+83.692320019 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.212445 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.212513 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.212532 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.212559 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.212578 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:11Z","lastTransitionTime":"2025-11-24T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.276292 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.276415 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.276427 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.276575 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.276616 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.276836 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.276950 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:11 crc kubenswrapper[4944]: E1124 08:53:11.277029 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.277896 4944 scope.go:117] "RemoveContainer" containerID="7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.315130 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.315629 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.315639 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.315659 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.315670 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:11Z","lastTransitionTime":"2025-11-24T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.418311 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.418350 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.418359 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.418375 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.418387 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:11Z","lastTransitionTime":"2025-11-24T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.520552 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.520602 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.520614 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.520634 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.520648 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:11Z","lastTransitionTime":"2025-11-24T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.627892 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.627939 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.627948 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.627966 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.627978 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:11Z","lastTransitionTime":"2025-11-24T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.635136 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/1.log" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.637442 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.637892 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.652609 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.716521 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.733003 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.733089 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.733102 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.733125 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.733138 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:11Z","lastTransitionTime":"2025-11-24T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.738863 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.756803 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.774422 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.788488 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.806122 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.825512 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.836072 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.836132 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.836144 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.836162 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.836172 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:11Z","lastTransitionTime":"2025-11-24T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.841255 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.857291 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.873018 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.888097 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.908454 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.928391 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.939582 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.939634 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.939644 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.939662 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.939674 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:11Z","lastTransitionTime":"2025-11-24T08:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.943870 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:11 crc kubenswrapper[4944]: I1124 08:53:11.963348 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:11Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.042335 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.042395 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.042407 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.042429 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.042444 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:12Z","lastTransitionTime":"2025-11-24T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.145088 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.145131 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.145143 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.145173 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.145191 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:12Z","lastTransitionTime":"2025-11-24T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.248616 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.248660 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.248673 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.248690 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.248702 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:12Z","lastTransitionTime":"2025-11-24T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.352097 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.352143 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.352154 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.352172 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.352185 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:12Z","lastTransitionTime":"2025-11-24T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.455648 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.455960 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.455984 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.456033 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.456091 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:12Z","lastTransitionTime":"2025-11-24T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.559521 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.559595 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.559619 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.559664 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.559702 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:12Z","lastTransitionTime":"2025-11-24T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.642870 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/2.log" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.643551 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/1.log" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.646962 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1" exitCode=1 Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.647033 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1"} Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.647177 4944 scope.go:117] "RemoveContainer" containerID="7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.647931 4944 scope.go:117] "RemoveContainer" containerID="ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1" Nov 24 08:53:12 crc kubenswrapper[4944]: E1124 08:53:12.648288 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.662580 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.662923 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.663107 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.663230 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.663314 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:12Z","lastTransitionTime":"2025-11-24T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.665366 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.682199 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.696887 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.711172 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.728010 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.741107 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.757437 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.766209 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.766309 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.766325 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.766350 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.766367 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:12Z","lastTransitionTime":"2025-11-24T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.769797 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.783418 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.797014 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.813467 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.836216 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2d90ad5c71fbf9d0e861d78d6e73e976b0ebf023ff96aac1cc1b2b0560ca1b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:12Z\\\",\\\"message\\\":\\\"ches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 08:53:12.129940 6626 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.849094 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.864022 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.868857 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.868898 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.868910 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.868932 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.868943 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:12Z","lastTransitionTime":"2025-11-24T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.879040 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.890774 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.972812 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.973321 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.973409 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.973622 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:12 crc kubenswrapper[4944]: I1124 08:53:12.973722 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:12Z","lastTransitionTime":"2025-11-24T08:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.077342 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.077397 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.077407 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.077428 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.077440 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:13Z","lastTransitionTime":"2025-11-24T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.180519 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.180567 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.180578 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.180598 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.180611 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:13Z","lastTransitionTime":"2025-11-24T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.275849 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.275960 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.275877 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:13 crc kubenswrapper[4944]: E1124 08:53:13.276023 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.276093 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:13 crc kubenswrapper[4944]: E1124 08:53:13.276197 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:13 crc kubenswrapper[4944]: E1124 08:53:13.276226 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:13 crc kubenswrapper[4944]: E1124 08:53:13.276290 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.283496 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.283544 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.283590 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.283613 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.283626 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:13Z","lastTransitionTime":"2025-11-24T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.386094 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.386169 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.386185 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.386206 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.386226 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:13Z","lastTransitionTime":"2025-11-24T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.488518 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.488569 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.488580 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.488599 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.488614 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:13Z","lastTransitionTime":"2025-11-24T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.591886 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.591939 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.591953 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.591977 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.591991 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:13Z","lastTransitionTime":"2025-11-24T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.653583 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/2.log" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.658759 4944 scope.go:117] "RemoveContainer" containerID="ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1" Nov 24 08:53:13 crc kubenswrapper[4944]: E1124 08:53:13.658962 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.672969 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.692609 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.701080 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.701131 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.701141 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.701159 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.701170 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:13Z","lastTransitionTime":"2025-11-24T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.720715 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:12Z\\\",\\\"message\\\":\\\"ches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 08:53:12.129940 6626 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.733922 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.750313 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.763671 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.777965 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.791615 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.804567 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.804622 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.804633 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.804652 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.804667 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:13Z","lastTransitionTime":"2025-11-24T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.814089 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.829400 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.843010 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.857683 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.878302 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.892399 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.906887 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.907954 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.907995 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.908005 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.908023 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.908035 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:13Z","lastTransitionTime":"2025-11-24T08:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:13 crc kubenswrapper[4944]: I1124 08:53:13.922169 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.011328 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.011380 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.011395 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.011414 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.011429 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:14Z","lastTransitionTime":"2025-11-24T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.114439 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.114496 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.114514 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.114535 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.114548 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:14Z","lastTransitionTime":"2025-11-24T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.216674 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.216762 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.216779 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.216800 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.216815 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:14Z","lastTransitionTime":"2025-11-24T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.319355 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.319417 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.319430 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.319446 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.319457 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:14Z","lastTransitionTime":"2025-11-24T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.421572 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.421662 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.421675 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.421694 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.421709 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:14Z","lastTransitionTime":"2025-11-24T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.524567 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.524618 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.524630 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.524648 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.524664 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:14Z","lastTransitionTime":"2025-11-24T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.628209 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.628261 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.628272 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.628291 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.628304 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:14Z","lastTransitionTime":"2025-11-24T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.731341 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.731405 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.731419 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.731440 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.731456 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:14Z","lastTransitionTime":"2025-11-24T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.835037 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.835145 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.835172 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.835206 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.835231 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:14Z","lastTransitionTime":"2025-11-24T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.938304 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.938368 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.938389 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.938419 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:14 crc kubenswrapper[4944]: I1124 08:53:14.938495 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:14Z","lastTransitionTime":"2025-11-24T08:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.041265 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.041324 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.041401 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.041427 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.041439 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:15Z","lastTransitionTime":"2025-11-24T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.144498 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.144546 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.144558 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.144575 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.144586 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:15Z","lastTransitionTime":"2025-11-24T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.248336 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.248400 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.248421 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.248580 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.249188 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:15Z","lastTransitionTime":"2025-11-24T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.275861 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:15 crc kubenswrapper[4944]: E1124 08:53:15.276117 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.276362 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.276434 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.276451 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:15 crc kubenswrapper[4944]: E1124 08:53:15.276871 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:15 crc kubenswrapper[4944]: E1124 08:53:15.277246 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:15 crc kubenswrapper[4944]: E1124 08:53:15.277470 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.352293 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.352342 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.352352 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.352370 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.352383 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:15Z","lastTransitionTime":"2025-11-24T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.455413 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.455461 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.455470 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.455486 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.455496 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:15Z","lastTransitionTime":"2025-11-24T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.558624 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.558993 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.559121 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.559249 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.559406 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:15Z","lastTransitionTime":"2025-11-24T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.662269 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.662333 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.662351 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.662378 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.662403 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:15Z","lastTransitionTime":"2025-11-24T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.766110 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.766166 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.766183 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.766210 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.766227 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:15Z","lastTransitionTime":"2025-11-24T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.869328 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.869412 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.869433 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.869456 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.869477 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:15Z","lastTransitionTime":"2025-11-24T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.973072 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.973126 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.973136 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.973162 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:15 crc kubenswrapper[4944]: I1124 08:53:15.973176 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:15Z","lastTransitionTime":"2025-11-24T08:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.076608 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.076697 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.076723 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.076757 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.076776 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.179156 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.179209 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.179221 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.179240 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.179254 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.282247 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.282315 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.282327 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.282349 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.282360 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.303496 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.317106 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.322639 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.339359 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.352126 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.367347 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.368579 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.368610 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.368622 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.368664 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.368700 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: E1124 08:53:16.380795 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.382214 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.385692 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.385746 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.385760 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.385784 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.385797 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: E1124 08:53:16.402218 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.404376 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:12Z\\\",\\\"message\\\":\\\"ches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 08:53:12.129940 6626 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.406850 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.406928 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.406959 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.406987 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.407000 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.417854 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: E1124 08:53:16.422574 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.426859 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.426919 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.426937 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.426958 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.426974 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.432734 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: E1124 08:53:16.440076 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.444742 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.444810 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.444824 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.444861 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.444890 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.448826 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: E1124 08:53:16.458914 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: E1124 08:53:16.459374 4944 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.461879 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.462265 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.462520 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.462829 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.463081 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.462189 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.477017 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.494494 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.511676 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.528677 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.543510 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.561448 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.566090 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.566130 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.566142 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.566161 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.566174 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.668666 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.668703 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.668714 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.668731 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.668740 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.771962 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.772008 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.772018 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.772035 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.772061 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.874923 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.874990 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.875016 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.875076 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.875093 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.977709 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.977772 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.977783 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.977801 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:16 crc kubenswrapper[4944]: I1124 08:53:16.977813 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:16Z","lastTransitionTime":"2025-11-24T08:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.081415 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.081496 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.081517 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.081546 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.081567 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:17Z","lastTransitionTime":"2025-11-24T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.184766 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.184808 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.184828 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.184856 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.184868 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:17Z","lastTransitionTime":"2025-11-24T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.276830 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.276939 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.276863 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.276893 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:17 crc kubenswrapper[4944]: E1124 08:53:17.277126 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:17 crc kubenswrapper[4944]: E1124 08:53:17.277258 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:17 crc kubenswrapper[4944]: E1124 08:53:17.277370 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:17 crc kubenswrapper[4944]: E1124 08:53:17.277456 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.287974 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.288024 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.288061 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.288080 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.288094 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:17Z","lastTransitionTime":"2025-11-24T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.390411 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.390471 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.390482 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.390505 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.390518 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:17Z","lastTransitionTime":"2025-11-24T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.493092 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.493189 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.493206 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.493223 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.493270 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:17Z","lastTransitionTime":"2025-11-24T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.596501 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.596557 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.596571 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.596590 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.596602 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:17Z","lastTransitionTime":"2025-11-24T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.699198 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.699268 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.699277 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.699315 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.699334 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:17Z","lastTransitionTime":"2025-11-24T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.802336 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.802423 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.802440 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.802461 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.802474 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:17Z","lastTransitionTime":"2025-11-24T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.905328 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.905396 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.905415 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.905441 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:17 crc kubenswrapper[4944]: I1124 08:53:17.905460 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:17Z","lastTransitionTime":"2025-11-24T08:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.008615 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.008733 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.008759 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.008795 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.008820 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:18Z","lastTransitionTime":"2025-11-24T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.111551 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.111646 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.111681 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.111714 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.111737 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:18Z","lastTransitionTime":"2025-11-24T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.214604 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.214649 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.214660 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.214676 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.214686 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:18Z","lastTransitionTime":"2025-11-24T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.318210 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.318265 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.318277 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.318298 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.318319 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:18Z","lastTransitionTime":"2025-11-24T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.421388 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.421479 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.421520 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.421552 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.421573 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:18Z","lastTransitionTime":"2025-11-24T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.525748 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.525824 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.525843 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.525870 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.525887 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:18Z","lastTransitionTime":"2025-11-24T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.629670 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.629731 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.629747 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.629769 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.629785 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:18Z","lastTransitionTime":"2025-11-24T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.732764 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.732844 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.732856 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.732878 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.732894 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:18Z","lastTransitionTime":"2025-11-24T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.835544 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.835623 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.835634 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.835653 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.835666 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:18Z","lastTransitionTime":"2025-11-24T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.939034 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.939096 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.939110 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.939129 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:18 crc kubenswrapper[4944]: I1124 08:53:18.939140 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:18Z","lastTransitionTime":"2025-11-24T08:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.041899 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.041946 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.041956 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.041973 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.041983 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:19Z","lastTransitionTime":"2025-11-24T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.145601 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.145648 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.145658 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.145674 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.145686 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:19Z","lastTransitionTime":"2025-11-24T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.248488 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.248541 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.248553 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.248573 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.248584 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:19Z","lastTransitionTime":"2025-11-24T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.276200 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.276294 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:19 crc kubenswrapper[4944]: E1124 08:53:19.276362 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.276318 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.276294 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:19 crc kubenswrapper[4944]: E1124 08:53:19.276465 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:19 crc kubenswrapper[4944]: E1124 08:53:19.276673 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:19 crc kubenswrapper[4944]: E1124 08:53:19.276815 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.352177 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.352230 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.352246 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.352272 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.352288 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:19Z","lastTransitionTime":"2025-11-24T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.454954 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.455729 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.455951 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.456042 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.456140 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:19Z","lastTransitionTime":"2025-11-24T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.558963 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.559024 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.559034 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.559072 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.559084 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:19Z","lastTransitionTime":"2025-11-24T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.661959 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.661999 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.662009 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.662027 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.662038 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:19Z","lastTransitionTime":"2025-11-24T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.765295 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.765348 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.765359 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.765376 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.765387 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:19Z","lastTransitionTime":"2025-11-24T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.868168 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.868206 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.868216 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.868233 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.868244 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:19Z","lastTransitionTime":"2025-11-24T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.970936 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.971002 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.971016 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.971079 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:19 crc kubenswrapper[4944]: I1124 08:53:19.971093 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:19Z","lastTransitionTime":"2025-11-24T08:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.074160 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.074539 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.074648 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.074762 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.074847 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:20Z","lastTransitionTime":"2025-11-24T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.178074 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.178126 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.178137 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.178152 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.178162 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:20Z","lastTransitionTime":"2025-11-24T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.281826 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.282253 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.282267 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.282286 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.282300 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:20Z","lastTransitionTime":"2025-11-24T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.292992 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.309298 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.322224 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.336107 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.349773 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.377210 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:12Z\\\",\\\"message\\\":\\\"ches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 08:53:12.129940 6626 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.386365 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.386400 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.386411 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.386429 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.386443 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:20Z","lastTransitionTime":"2025-11-24T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.389792 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.404884 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.421804 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.436807 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.453990 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bebeee-67e1-41eb-902f-863da7195b46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://739cd0cec0794b22aa306490236c6e42aef2aab4b8787cfed8752e727d1c48c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10a53461679aed9575532a08cb82b6bae353bf170c7a663369aeb4dac512e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94d33673d9208d6570c2853f6644328f2997815a8bfd48a5d7338d12bab2f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.467869 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.482221 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.488871 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.488902 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.488912 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.488927 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.488936 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:20Z","lastTransitionTime":"2025-11-24T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.496540 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.514425 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.530066 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.544152 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:20Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.590960 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.591042 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.591090 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.591116 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.591133 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:20Z","lastTransitionTime":"2025-11-24T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.694118 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.694169 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.694182 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.694204 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.694217 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:20Z","lastTransitionTime":"2025-11-24T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.796992 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.797040 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.797066 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.797084 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.797097 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:20Z","lastTransitionTime":"2025-11-24T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.900242 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.900300 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.900318 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.900345 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:20 crc kubenswrapper[4944]: I1124 08:53:20.900364 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:20Z","lastTransitionTime":"2025-11-24T08:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.003766 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.003812 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.003829 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.003852 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.003868 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:21Z","lastTransitionTime":"2025-11-24T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.106314 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.106720 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.106798 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.106886 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.106958 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:21Z","lastTransitionTime":"2025-11-24T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.210338 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.210382 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.210395 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.210413 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.210426 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:21Z","lastTransitionTime":"2025-11-24T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.276215 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.276327 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:21 crc kubenswrapper[4944]: E1124 08:53:21.276721 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.276409 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:21 crc kubenswrapper[4944]: E1124 08:53:21.276888 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.276346 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:21 crc kubenswrapper[4944]: E1124 08:53:21.277100 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:21 crc kubenswrapper[4944]: E1124 08:53:21.277369 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.313505 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.313555 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.313567 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.313586 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.313598 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:21Z","lastTransitionTime":"2025-11-24T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.416704 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.416739 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.416750 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.416765 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.416775 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:21Z","lastTransitionTime":"2025-11-24T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.519785 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.520538 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.520638 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.520749 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.520836 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:21Z","lastTransitionTime":"2025-11-24T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.624375 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.624412 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.624420 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.624435 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.624445 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:21Z","lastTransitionTime":"2025-11-24T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.728087 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.728274 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.728292 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.728320 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.728344 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:21Z","lastTransitionTime":"2025-11-24T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.831388 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.831453 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.831467 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.831484 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.831497 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:21Z","lastTransitionTime":"2025-11-24T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.934948 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.935007 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.935019 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.935041 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:21 crc kubenswrapper[4944]: I1124 08:53:21.935075 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:21Z","lastTransitionTime":"2025-11-24T08:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.037758 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.037837 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.037846 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.037865 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.037877 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:22Z","lastTransitionTime":"2025-11-24T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.141183 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.141233 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.141248 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.141270 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.141285 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:22Z","lastTransitionTime":"2025-11-24T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.243869 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.243925 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.243936 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.243957 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.243970 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:22Z","lastTransitionTime":"2025-11-24T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.346757 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.346806 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.346816 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.346831 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.346843 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:22Z","lastTransitionTime":"2025-11-24T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.450095 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.450138 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.450148 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.450172 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.450192 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:22Z","lastTransitionTime":"2025-11-24T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.553693 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.553744 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.553759 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.554457 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.554484 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:22Z","lastTransitionTime":"2025-11-24T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.657560 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.657619 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.657635 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.657659 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.657678 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:22Z","lastTransitionTime":"2025-11-24T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.760400 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.760479 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.760491 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.760512 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.760528 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:22Z","lastTransitionTime":"2025-11-24T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.863340 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.863376 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.863385 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.863404 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.863417 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:22Z","lastTransitionTime":"2025-11-24T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.966812 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.966856 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.966869 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.966886 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:22 crc kubenswrapper[4944]: I1124 08:53:22.966896 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:22Z","lastTransitionTime":"2025-11-24T08:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.070450 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.070501 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.070516 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.070537 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.070552 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:23Z","lastTransitionTime":"2025-11-24T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.172952 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.173016 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.173027 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.173061 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.173076 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:23Z","lastTransitionTime":"2025-11-24T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.275738 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.275738 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.275766 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.275841 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.275859 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.275835 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.275880 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.275898 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:23Z","lastTransitionTime":"2025-11-24T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.275914 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:23 crc kubenswrapper[4944]: E1124 08:53:23.276181 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:23 crc kubenswrapper[4944]: E1124 08:53:23.276279 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:23 crc kubenswrapper[4944]: E1124 08:53:23.276329 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:23 crc kubenswrapper[4944]: E1124 08:53:23.276461 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.378984 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.379031 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.379074 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.379096 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.379109 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:23Z","lastTransitionTime":"2025-11-24T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.481840 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.482003 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.482014 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.482032 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.482063 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:23Z","lastTransitionTime":"2025-11-24T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.584837 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.584911 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.584928 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.584951 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.584962 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:23Z","lastTransitionTime":"2025-11-24T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.688407 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.688488 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.688503 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.688525 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.688539 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:23Z","lastTransitionTime":"2025-11-24T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.791787 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.791860 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.791874 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.791895 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.791911 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:23Z","lastTransitionTime":"2025-11-24T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.894941 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.895003 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.895014 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.895034 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.895082 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:23Z","lastTransitionTime":"2025-11-24T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.997886 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.997929 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.997939 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.997962 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:23 crc kubenswrapper[4944]: I1124 08:53:23.997976 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:23Z","lastTransitionTime":"2025-11-24T08:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.100326 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.100370 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.100378 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.100400 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.100412 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:24Z","lastTransitionTime":"2025-11-24T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.203147 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.203236 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.203251 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.203269 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.203280 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:24Z","lastTransitionTime":"2025-11-24T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.306398 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.306451 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.306461 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.306482 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.306495 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:24Z","lastTransitionTime":"2025-11-24T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.409432 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.409478 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.409488 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.409507 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.409520 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:24Z","lastTransitionTime":"2025-11-24T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.512240 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.512276 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.512285 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.512299 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.512311 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:24Z","lastTransitionTime":"2025-11-24T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.615538 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.615600 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.615611 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.615633 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.615646 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:24Z","lastTransitionTime":"2025-11-24T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.717795 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.717833 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.717843 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.717859 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.717877 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:24Z","lastTransitionTime":"2025-11-24T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.819839 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.819882 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.819892 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.819908 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.819918 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:24Z","lastTransitionTime":"2025-11-24T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.922608 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.922681 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.922695 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.922719 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:24 crc kubenswrapper[4944]: I1124 08:53:24.922732 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:24Z","lastTransitionTime":"2025-11-24T08:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.025056 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.025110 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.025125 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.025147 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.025159 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:25Z","lastTransitionTime":"2025-11-24T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.128488 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.128545 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.128562 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.128583 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.128594 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:25Z","lastTransitionTime":"2025-11-24T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.231387 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.231442 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.231457 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.231476 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.231488 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:25Z","lastTransitionTime":"2025-11-24T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.276268 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.276388 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:25 crc kubenswrapper[4944]: E1124 08:53:25.276449 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.276519 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.276621 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:25 crc kubenswrapper[4944]: E1124 08:53:25.276683 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:25 crc kubenswrapper[4944]: E1124 08:53:25.276819 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:25 crc kubenswrapper[4944]: E1124 08:53:25.276931 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.334689 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.334733 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.334743 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.334760 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.334771 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:25Z","lastTransitionTime":"2025-11-24T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.437925 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.438733 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.438762 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.438814 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.438833 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:25Z","lastTransitionTime":"2025-11-24T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.523750 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:25 crc kubenswrapper[4944]: E1124 08:53:25.523889 4944 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:53:25 crc kubenswrapper[4944]: E1124 08:53:25.523959 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs podName:55990e0c-a6ff-4b18-8b8d-f27542761408 nodeName:}" failed. No retries permitted until 2025-11-24 08:53:57.523943281 +0000 UTC m=+98.058383743 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs") pod "network-metrics-daemon-jmkb9" (UID: "55990e0c-a6ff-4b18-8b8d-f27542761408") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.542526 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.542583 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.542593 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.542613 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.542631 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:25Z","lastTransitionTime":"2025-11-24T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.645397 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.645448 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.645461 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.645478 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.645490 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:25Z","lastTransitionTime":"2025-11-24T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.748190 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.748231 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.748241 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.748258 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.748271 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:25Z","lastTransitionTime":"2025-11-24T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.851546 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.851586 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.851598 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.851627 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.851643 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:25Z","lastTransitionTime":"2025-11-24T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.954101 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.954163 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.954177 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.954203 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:25 crc kubenswrapper[4944]: I1124 08:53:25.954224 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:25Z","lastTransitionTime":"2025-11-24T08:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.057233 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.057291 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.057304 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.057327 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.057342 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.160191 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.160235 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.160244 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.160259 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.160271 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.262800 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.262845 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.262856 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.262875 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.262887 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.365907 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.365943 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.365956 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.365973 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.365984 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.468491 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.468538 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.468549 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.468577 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.468590 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.570837 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.570899 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.570910 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.570931 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.570947 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.674039 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.674106 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.674117 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.674136 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.674150 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.686615 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.686688 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.686804 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.686849 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.686865 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: E1124 08:53:26.701227 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.706017 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.706086 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.706100 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.706124 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.706140 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: E1124 08:53:26.722741 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.727924 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.727981 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.727995 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.728017 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.728036 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: E1124 08:53:26.750762 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.755930 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.755982 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.755996 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.756014 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.756025 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: E1124 08:53:26.772204 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.776424 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.776472 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.776486 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.776506 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.776519 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: E1124 08:53:26.791647 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:26 crc kubenswrapper[4944]: E1124 08:53:26.791818 4944 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.793974 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.794021 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.794032 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.794067 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.794085 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.898229 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.898569 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.898668 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.898777 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:26 crc kubenswrapper[4944]: I1124 08:53:26.898861 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:26Z","lastTransitionTime":"2025-11-24T08:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.001651 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.001693 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.001709 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.001732 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.001743 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:27Z","lastTransitionTime":"2025-11-24T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.104848 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.104905 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.104917 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.104932 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.104943 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:27Z","lastTransitionTime":"2025-11-24T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.207976 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.208024 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.208037 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.208081 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.208096 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:27Z","lastTransitionTime":"2025-11-24T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.276301 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.276405 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.276444 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:27 crc kubenswrapper[4944]: E1124 08:53:27.276511 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.276344 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:27 crc kubenswrapper[4944]: E1124 08:53:27.276664 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:27 crc kubenswrapper[4944]: E1124 08:53:27.276846 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:27 crc kubenswrapper[4944]: E1124 08:53:27.277012 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.310654 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.311005 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.311165 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.311281 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.311395 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:27Z","lastTransitionTime":"2025-11-24T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.414913 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.414962 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.414975 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.414994 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.415007 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:27Z","lastTransitionTime":"2025-11-24T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.519397 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.519900 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.520157 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.520390 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.520586 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:27Z","lastTransitionTime":"2025-11-24T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.623371 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.623436 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.623474 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.623497 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.623512 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:27Z","lastTransitionTime":"2025-11-24T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.725999 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.726060 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.726073 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.726090 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.726103 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:27Z","lastTransitionTime":"2025-11-24T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.833541 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.833586 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.833598 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.833618 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.833629 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:27Z","lastTransitionTime":"2025-11-24T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.936790 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.936852 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.936874 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.936900 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:27 crc kubenswrapper[4944]: I1124 08:53:27.936919 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:27Z","lastTransitionTime":"2025-11-24T08:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.041307 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.041356 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.041369 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.041394 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.041413 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:28Z","lastTransitionTime":"2025-11-24T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.145456 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.145507 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.145519 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.145542 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.145555 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:28Z","lastTransitionTime":"2025-11-24T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.248557 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.248605 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.248619 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.248637 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.248651 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:28Z","lastTransitionTime":"2025-11-24T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.351647 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.351712 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.351725 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.351747 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.351765 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:28Z","lastTransitionTime":"2025-11-24T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.458757 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.459201 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.459315 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.459394 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.459472 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:28Z","lastTransitionTime":"2025-11-24T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.561850 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.561894 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.561904 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.561921 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.561932 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:28Z","lastTransitionTime":"2025-11-24T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.664342 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.664394 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.664405 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.664421 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.664432 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:28Z","lastTransitionTime":"2025-11-24T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.712021 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dg4dk_5ff38b2b-032d-47df-8836-105e0cfae835/kube-multus/0.log" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.712127 4944 generic.go:334] "Generic (PLEG): container finished" podID="5ff38b2b-032d-47df-8836-105e0cfae835" containerID="d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6" exitCode=1 Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.712169 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dg4dk" event={"ID":"5ff38b2b-032d-47df-8836-105e0cfae835","Type":"ContainerDied","Data":"d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.712736 4944 scope.go:117] "RemoveContainer" containerID="d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.738657 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.757072 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:28Z\\\",\\\"message\\\":\\\"2025-11-24T08:52:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5\\\\n2025-11-24T08:52:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5 to /host/opt/cni/bin/\\\\n2025-11-24T08:52:43Z [verbose] multus-daemon started\\\\n2025-11-24T08:52:43Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:53:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.767755 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.767792 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.767803 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.767823 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.767838 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:28Z","lastTransitionTime":"2025-11-24T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.781221 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:12Z\\\",\\\"message\\\":\\\"ches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 08:53:12.129940 6626 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.794062 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.813950 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.831497 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.845420 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.861088 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.870974 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.871241 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.871379 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.871496 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.871616 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:28Z","lastTransitionTime":"2025-11-24T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.881919 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bebeee-67e1-41eb-902f-863da7195b46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://739cd0cec0794b22aa306490236c6e42aef2aab4b8787cfed8752e727d1c48c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10a53461679aed9575532a08cb82b6bae353bf170c7a663369aeb4dac512e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94d33673d9208d6570c2853f6644328f2997815a8bfd48a5d7338d12bab2f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.901658 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.920022 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.937625 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.956404 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.975743 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.976184 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.976207 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.976219 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.976234 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.976246 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:28Z","lastTransitionTime":"2025-11-24T08:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:28 crc kubenswrapper[4944]: I1124 08:53:28.989844 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:28Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.005654 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.021345 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.078789 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.078833 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.078843 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.078861 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.078874 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:29Z","lastTransitionTime":"2025-11-24T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.181971 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.182188 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.182278 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.182364 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.182450 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:29Z","lastTransitionTime":"2025-11-24T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.276589 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.276918 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.276944 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.276994 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.277246 4944 scope.go:117] "RemoveContainer" containerID="ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1" Nov 24 08:53:29 crc kubenswrapper[4944]: E1124 08:53:29.277239 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:29 crc kubenswrapper[4944]: E1124 08:53:29.277398 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:29 crc kubenswrapper[4944]: E1124 08:53:29.277420 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" Nov 24 08:53:29 crc kubenswrapper[4944]: E1124 08:53:29.278186 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:29 crc kubenswrapper[4944]: E1124 08:53:29.278307 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.286039 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.286093 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.286110 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.286126 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.286137 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:29Z","lastTransitionTime":"2025-11-24T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.308150 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.389500 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.389547 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.389559 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.389581 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.389595 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:29Z","lastTransitionTime":"2025-11-24T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.492855 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.492901 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.492910 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.492931 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.492941 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:29Z","lastTransitionTime":"2025-11-24T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.595331 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.595380 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.595394 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.595410 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.595425 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:29Z","lastTransitionTime":"2025-11-24T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.698695 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.698753 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.698763 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.698780 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.698791 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:29Z","lastTransitionTime":"2025-11-24T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.723323 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dg4dk_5ff38b2b-032d-47df-8836-105e0cfae835/kube-multus/0.log" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.723445 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dg4dk" event={"ID":"5ff38b2b-032d-47df-8836-105e0cfae835","Type":"ContainerStarted","Data":"df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f"} Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.740609 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.758652 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.776748 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.793251 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.802251 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.802328 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.802345 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.802367 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.802382 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:29Z","lastTransitionTime":"2025-11-24T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.807806 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6043c3-7c9c-4773-b43a-51026a0d0ff5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae6958eb78d429d6c25449d6b5595b7ef19937a8db223b185317ae344b8dd947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.823001 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.839018 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:28Z\\\",\\\"message\\\":\\\"2025-11-24T08:52:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5\\\\n2025-11-24T08:52:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5 to /host/opt/cni/bin/\\\\n2025-11-24T08:52:43Z [verbose] multus-daemon started\\\\n2025-11-24T08:52:43Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:53:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:53:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.861489 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:12Z\\\",\\\"message\\\":\\\"ches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 08:53:12.129940 6626 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.874203 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.889197 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bebeee-67e1-41eb-902f-863da7195b46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://739cd0cec0794b22aa306490236c6e42aef2aab4b8787cfed8752e727d1c48c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10a53461679aed9575532a08cb82b6bae353bf170c7a663369aeb4dac512e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94d33673d9208d6570c2853f6644328f2997815a8bfd48a5d7338d12bab2f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.904805 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.906109 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.906156 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.906168 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.906187 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.906213 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:29Z","lastTransitionTime":"2025-11-24T08:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.919080 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.934393 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.948730 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.964239 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.976664 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:29 crc kubenswrapper[4944]: I1124 08:53:29.991270 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:29Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.008669 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.009833 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.009888 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.009900 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.009926 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.009939 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:30Z","lastTransitionTime":"2025-11-24T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.113266 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.113314 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.113324 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.113343 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.113353 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:30Z","lastTransitionTime":"2025-11-24T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.215890 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.215953 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.215968 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.215988 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.216004 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:30Z","lastTransitionTime":"2025-11-24T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.290525 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.305418 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.319287 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.319596 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.319704 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.319802 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.319909 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:30Z","lastTransitionTime":"2025-11-24T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.320016 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6043c3-7c9c-4773-b43a-51026a0d0ff5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae6958eb78d429d6c25449d6b5595b7ef19937a8db223b185317ae344b8dd947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.340822 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.357678 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:28Z\\\",\\\"message\\\":\\\"2025-11-24T08:52:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5\\\\n2025-11-24T08:52:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5 to /host/opt/cni/bin/\\\\n2025-11-24T08:52:43Z [verbose] multus-daemon started\\\\n2025-11-24T08:52:43Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:53:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:53:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.379267 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:12Z\\\",\\\"message\\\":\\\"ches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 08:53:12.129940 6626 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.393974 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bebeee-67e1-41eb-902f-863da7195b46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://739cd0cec0794b22aa306490236c6e42aef2aab4b8787cfed8752e727d1c48c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10a53461679aed9575532a08cb82b6bae353bf170c7a663369aeb4dac512e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94d33673d9208d6570c2853f6644328f2997815a8bfd48a5d7338d12bab2f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.410560 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.422358 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.422685 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.422737 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.422749 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.422769 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.422781 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:30Z","lastTransitionTime":"2025-11-24T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.435506 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.450659 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.464147 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.479962 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.496507 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.511178 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.525943 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.526001 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.526012 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.526031 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.526387 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:30Z","lastTransitionTime":"2025-11-24T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.531345 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.544125 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.556461 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:30Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.631730 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.631793 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.631807 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.631923 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.631938 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:30Z","lastTransitionTime":"2025-11-24T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.734456 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.734523 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.734542 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.734562 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.734574 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:30Z","lastTransitionTime":"2025-11-24T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.838677 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.838735 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.838747 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.838770 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.838783 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:30Z","lastTransitionTime":"2025-11-24T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.941521 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.941573 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.941582 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.941599 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:30 crc kubenswrapper[4944]: I1124 08:53:30.941610 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:30Z","lastTransitionTime":"2025-11-24T08:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.043916 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.043961 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.043974 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.043994 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.044005 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:31Z","lastTransitionTime":"2025-11-24T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.146812 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.146847 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.146856 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.146871 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.146881 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:31Z","lastTransitionTime":"2025-11-24T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.249495 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.249542 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.249553 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.249571 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.249585 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:31Z","lastTransitionTime":"2025-11-24T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.275971 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.276036 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.276157 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:31 crc kubenswrapper[4944]: E1124 08:53:31.276153 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.276203 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:31 crc kubenswrapper[4944]: E1124 08:53:31.276246 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:31 crc kubenswrapper[4944]: E1124 08:53:31.276339 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:31 crc kubenswrapper[4944]: E1124 08:53:31.276482 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.352198 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.352235 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.352247 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.352264 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.352275 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:31Z","lastTransitionTime":"2025-11-24T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.454986 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.455041 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.455065 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.455082 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.455094 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:31Z","lastTransitionTime":"2025-11-24T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.558122 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.558165 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.558194 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.558219 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.558236 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:31Z","lastTransitionTime":"2025-11-24T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.660761 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.660799 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.660809 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.660828 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.660847 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:31Z","lastTransitionTime":"2025-11-24T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.763225 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.763612 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.763696 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.763819 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.763879 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:31Z","lastTransitionTime":"2025-11-24T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.867055 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.867103 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.867113 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.867132 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.867159 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:31Z","lastTransitionTime":"2025-11-24T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.969127 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.969175 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.969186 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.969204 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:31 crc kubenswrapper[4944]: I1124 08:53:31.969214 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:31Z","lastTransitionTime":"2025-11-24T08:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.072197 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.072253 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.072263 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.072284 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.072297 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:32Z","lastTransitionTime":"2025-11-24T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.174957 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.175427 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.175528 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.175619 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.175692 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:32Z","lastTransitionTime":"2025-11-24T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.280456 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.280755 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.280830 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.280911 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.280987 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:32Z","lastTransitionTime":"2025-11-24T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.383458 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.383504 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.383514 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.383724 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.383741 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:32Z","lastTransitionTime":"2025-11-24T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.487083 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.487133 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.487144 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.487161 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.487175 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:32Z","lastTransitionTime":"2025-11-24T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.589957 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.589999 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.590012 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.590030 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.590064 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:32Z","lastTransitionTime":"2025-11-24T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.692272 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.692326 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.692336 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.692356 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.692368 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:32Z","lastTransitionTime":"2025-11-24T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.794975 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.795022 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.795035 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.795067 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.795083 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:32Z","lastTransitionTime":"2025-11-24T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.897952 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.898034 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.898083 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.898129 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:32 crc kubenswrapper[4944]: I1124 08:53:32.898148 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:32Z","lastTransitionTime":"2025-11-24T08:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.000445 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.000501 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.000515 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.000539 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.000555 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:33Z","lastTransitionTime":"2025-11-24T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.103783 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.103829 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.103837 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.103853 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.103864 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:33Z","lastTransitionTime":"2025-11-24T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.206760 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.206808 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.206827 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.206869 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.206891 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:33Z","lastTransitionTime":"2025-11-24T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.276799 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.276894 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.276949 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:33 crc kubenswrapper[4944]: E1124 08:53:33.277200 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:33 crc kubenswrapper[4944]: E1124 08:53:33.277321 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.277340 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:33 crc kubenswrapper[4944]: E1124 08:53:33.277418 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:33 crc kubenswrapper[4944]: E1124 08:53:33.277486 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.309978 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.310027 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.310038 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.310112 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.310126 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:33Z","lastTransitionTime":"2025-11-24T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.414814 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.414861 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.414872 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.414889 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.414899 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:33Z","lastTransitionTime":"2025-11-24T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.517371 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.517430 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.517441 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.517458 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.517473 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:33Z","lastTransitionTime":"2025-11-24T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.620878 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.620917 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.620929 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.620947 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.620960 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:33Z","lastTransitionTime":"2025-11-24T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.722909 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.722953 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.722963 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.722978 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.722989 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:33Z","lastTransitionTime":"2025-11-24T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.826582 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.826646 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.826668 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.826694 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.826712 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:33Z","lastTransitionTime":"2025-11-24T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.930028 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.930104 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.930115 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.930136 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:33 crc kubenswrapper[4944]: I1124 08:53:33.930151 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:33Z","lastTransitionTime":"2025-11-24T08:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.032895 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.033258 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.033335 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.033429 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.033499 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:34Z","lastTransitionTime":"2025-11-24T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.136000 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.136063 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.136076 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.136098 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.136113 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:34Z","lastTransitionTime":"2025-11-24T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.238486 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.238927 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.239002 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.239102 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.239180 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:34Z","lastTransitionTime":"2025-11-24T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.342429 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.342969 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.343193 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.343365 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.343506 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:34Z","lastTransitionTime":"2025-11-24T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.446819 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.446862 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.446871 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.446888 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.446899 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:34Z","lastTransitionTime":"2025-11-24T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.549547 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.549582 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.549591 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.549606 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.549616 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:34Z","lastTransitionTime":"2025-11-24T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.652286 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.652745 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.652881 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.652985 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.653104 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:34Z","lastTransitionTime":"2025-11-24T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.756349 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.756415 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.756426 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.756445 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.756456 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:34Z","lastTransitionTime":"2025-11-24T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.859310 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.859357 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.859366 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.859385 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.859394 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:34Z","lastTransitionTime":"2025-11-24T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.961989 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.962029 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.962038 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.962060 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:34 crc kubenswrapper[4944]: I1124 08:53:34.962081 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:34Z","lastTransitionTime":"2025-11-24T08:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.064237 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.064279 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.064288 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.064303 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.064313 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:35Z","lastTransitionTime":"2025-11-24T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.167556 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.167603 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.167614 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.167633 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.167646 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:35Z","lastTransitionTime":"2025-11-24T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.269988 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.270048 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.270089 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.270110 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.270125 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:35Z","lastTransitionTime":"2025-11-24T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.276440 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.276450 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:35 crc kubenswrapper[4944]: E1124 08:53:35.276588 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.276450 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:35 crc kubenswrapper[4944]: E1124 08:53:35.276668 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.276450 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:35 crc kubenswrapper[4944]: E1124 08:53:35.276746 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:35 crc kubenswrapper[4944]: E1124 08:53:35.276820 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.373035 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.373107 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.373120 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.373159 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.373175 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:35Z","lastTransitionTime":"2025-11-24T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.475223 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.475264 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.475272 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.475286 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.475297 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:35Z","lastTransitionTime":"2025-11-24T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.578983 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.579035 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.579046 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.579081 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.579093 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:35Z","lastTransitionTime":"2025-11-24T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.681818 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.681865 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.681879 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.681898 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.681914 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:35Z","lastTransitionTime":"2025-11-24T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.786196 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.786248 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.786258 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.786276 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.786287 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:35Z","lastTransitionTime":"2025-11-24T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.889092 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.889127 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.889136 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.889151 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.889160 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:35Z","lastTransitionTime":"2025-11-24T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.992571 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.992633 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.992645 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.992709 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:35 crc kubenswrapper[4944]: I1124 08:53:35.992722 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:35Z","lastTransitionTime":"2025-11-24T08:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.095300 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.095349 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.095359 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.095374 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.095385 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:36Z","lastTransitionTime":"2025-11-24T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.197943 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.197996 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.198007 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.198025 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.198037 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:36Z","lastTransitionTime":"2025-11-24T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.300161 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.300202 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.300213 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.300264 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.300280 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:36Z","lastTransitionTime":"2025-11-24T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.403438 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.403489 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.403500 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.403517 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.403528 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:36Z","lastTransitionTime":"2025-11-24T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.505809 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.505846 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.505855 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.505871 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.505882 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:36Z","lastTransitionTime":"2025-11-24T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.608692 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.608746 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.608763 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.608785 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.608797 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:36Z","lastTransitionTime":"2025-11-24T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.711738 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.711770 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.711779 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.711794 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.711809 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:36Z","lastTransitionTime":"2025-11-24T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.814108 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.814444 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.814541 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.814627 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.814688 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:36Z","lastTransitionTime":"2025-11-24T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.920805 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.920845 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.920855 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.920870 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.920879 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:36Z","lastTransitionTime":"2025-11-24T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.988877 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.988952 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.988962 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.988980 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:36 crc kubenswrapper[4944]: I1124 08:53:36.988991 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:36Z","lastTransitionTime":"2025-11-24T08:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: E1124 08:53:37.002292 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.006253 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.006289 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.006308 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.006329 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.006342 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: E1124 08:53:37.017565 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.020642 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.020672 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.020681 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.020696 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.020706 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: E1124 08:53:37.031464 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.034900 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.034938 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.034947 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.034962 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.034973 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: E1124 08:53:37.046236 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.049912 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.049956 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.049967 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.049985 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.049996 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: E1124 08:53:37.061297 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:37 crc kubenswrapper[4944]: E1124 08:53:37.061416 4944 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.063272 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.063309 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.063318 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.063333 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.063346 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.166486 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.166530 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.166542 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.166560 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.166572 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.269029 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.269086 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.269096 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.269110 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.269120 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.276505 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.276526 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.276569 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.276526 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:37 crc kubenswrapper[4944]: E1124 08:53:37.276642 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:37 crc kubenswrapper[4944]: E1124 08:53:37.276764 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:37 crc kubenswrapper[4944]: E1124 08:53:37.276863 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:37 crc kubenswrapper[4944]: E1124 08:53:37.276921 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.371559 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.371601 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.371612 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.371629 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.371642 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.474556 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.474596 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.474606 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.474623 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.474636 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.577488 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.577991 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.578171 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.578302 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.578384 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.681851 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.681884 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.681892 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.681909 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.681921 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.784947 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.785047 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.785087 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.785106 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.785121 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.890624 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.890701 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.890720 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.890763 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.890780 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.993904 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.993970 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.993984 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.994003 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:37 crc kubenswrapper[4944]: I1124 08:53:37.994016 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:37Z","lastTransitionTime":"2025-11-24T08:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.097314 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.097356 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.097365 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.097389 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.097399 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:38Z","lastTransitionTime":"2025-11-24T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.199639 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.199697 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.199712 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.199736 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.199750 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:38Z","lastTransitionTime":"2025-11-24T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.302796 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.302834 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.302844 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.302902 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.302918 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:38Z","lastTransitionTime":"2025-11-24T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.406198 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.406241 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.406249 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.406266 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.406275 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:38Z","lastTransitionTime":"2025-11-24T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.508529 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.509108 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.509120 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.509142 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.509151 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:38Z","lastTransitionTime":"2025-11-24T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.610950 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.610987 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.610995 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.611012 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.611021 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:38Z","lastTransitionTime":"2025-11-24T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.714876 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.714913 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.714923 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.714940 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.714951 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:38Z","lastTransitionTime":"2025-11-24T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.818210 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.818262 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.818274 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.818305 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.818318 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:38Z","lastTransitionTime":"2025-11-24T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.921975 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.922027 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.922039 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.922084 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:38 crc kubenswrapper[4944]: I1124 08:53:38.922095 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:38Z","lastTransitionTime":"2025-11-24T08:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.025364 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.025420 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.025433 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.025450 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.025463 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:39Z","lastTransitionTime":"2025-11-24T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.127751 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.127786 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.127794 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.127810 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.127819 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:39Z","lastTransitionTime":"2025-11-24T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.230987 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.231393 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.231625 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.231835 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.232037 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:39Z","lastTransitionTime":"2025-11-24T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.276293 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:39 crc kubenswrapper[4944]: E1124 08:53:39.276453 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.276688 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:39 crc kubenswrapper[4944]: E1124 08:53:39.276757 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.276937 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.277100 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:39 crc kubenswrapper[4944]: E1124 08:53:39.277175 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:39 crc kubenswrapper[4944]: E1124 08:53:39.277416 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.335056 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.335096 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.335110 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.335130 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.335143 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:39Z","lastTransitionTime":"2025-11-24T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.437458 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.437505 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.437516 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.437544 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.437556 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:39Z","lastTransitionTime":"2025-11-24T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.540525 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.540638 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.540650 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.540665 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.540675 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:39Z","lastTransitionTime":"2025-11-24T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.643708 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.643756 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.643766 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.643781 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.643792 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:39Z","lastTransitionTime":"2025-11-24T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.746799 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.746842 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.746853 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.746876 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.746889 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:39Z","lastTransitionTime":"2025-11-24T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.849907 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.849971 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.849984 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.850005 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.850022 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:39Z","lastTransitionTime":"2025-11-24T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.952396 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.952760 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.952845 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.952917 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:39 crc kubenswrapper[4944]: I1124 08:53:39.952980 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:39Z","lastTransitionTime":"2025-11-24T08:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.055300 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.055338 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.055347 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.055363 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.055377 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:40Z","lastTransitionTime":"2025-11-24T08:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.157919 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.157952 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.157962 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.157977 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.157989 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:40Z","lastTransitionTime":"2025-11-24T08:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.260104 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.260454 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.260540 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.260778 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.260854 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:40Z","lastTransitionTime":"2025-11-24T08:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.277498 4944 scope.go:117] "RemoveContainer" containerID="ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.293108 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6043c3-7c9c-4773-b43a-51026a0d0ff5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae6958eb78d429d6c25449d6b5595b7ef19937a8db223b185317ae344b8dd947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.306271 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.319389 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:28Z\\\",\\\"message\\\":\\\"2025-11-24T08:52:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5\\\\n2025-11-24T08:52:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5 to /host/opt/cni/bin/\\\\n2025-11-24T08:52:43Z [verbose] multus-daemon started\\\\n2025-11-24T08:52:43Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:53:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:53:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.339346 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:12Z\\\",\\\"message\\\":\\\"ches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 08:53:12.129940 6626 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.351993 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.363479 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.363519 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.363528 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.363545 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.363556 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:40Z","lastTransitionTime":"2025-11-24T08:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.367025 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.382489 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bebeee-67e1-41eb-902f-863da7195b46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://739cd0cec0794b22aa306490236c6e42aef2aab4b8787cfed8752e727d1c48c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10a53461679aed9575532a08cb82b6bae353bf170c7a663369aeb4dac512e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94d33673d9208d6570c2853f6644328f2997815a8bfd48a5d7338d12bab2f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.399991 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.414112 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.427670 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.444691 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.458697 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.473636 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.473684 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.473696 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.473712 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.473665 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.473723 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:40Z","lastTransitionTime":"2025-11-24T08:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.487499 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.505984 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.521647 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.535853 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.549604 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.577343 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.577385 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.577394 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.577410 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.577420 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:40Z","lastTransitionTime":"2025-11-24T08:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.679755 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.679802 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.679814 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.679832 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.679846 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:40Z","lastTransitionTime":"2025-11-24T08:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.762970 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/2.log" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.765364 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.765855 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.778940 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.782647 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.782699 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.782711 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.782730 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.782748 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:40Z","lastTransitionTime":"2025-11-24T08:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.792343 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bebeee-67e1-41eb-902f-863da7195b46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://739cd0cec0794b22aa306490236c6e42aef2aab4b8787cfed8752e727d1c48c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10a53461679aed9575532a08cb82b6bae353bf170c7a663369aeb4dac512e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94d33673d9208d6570c2853f6644328f2997815a8bfd48a5d7338d12bab2f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.805734 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.817817 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.831358 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.850108 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.871564 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.883808 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.885347 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.885380 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.885393 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.885409 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.885420 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:40Z","lastTransitionTime":"2025-11-24T08:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.897317 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.908263 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.919334 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.930419 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.942018 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:28Z\\\",\\\"message\\\":\\\"2025-11-24T08:52:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5\\\\n2025-11-24T08:52:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5 to /host/opt/cni/bin/\\\\n2025-11-24T08:52:43Z [verbose] multus-daemon started\\\\n2025-11-24T08:52:43Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:53:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:53:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.958491 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:12Z\\\",\\\"message\\\":\\\"ches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 08:53:12.129940 6626 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:53:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.969095 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.979689 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.988307 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.988348 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.988361 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.988381 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.988432 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:40Z","lastTransitionTime":"2025-11-24T08:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:40 crc kubenswrapper[4944]: I1124 08:53:40.989824 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6043c3-7c9c-4773-b43a-51026a0d0ff5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae6958eb78d429d6c25449d6b5595b7ef19937a8db223b185317ae344b8dd947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:40Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.005510 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.090475 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.090535 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.090545 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.090563 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.090575 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:41Z","lastTransitionTime":"2025-11-24T08:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.193151 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.193203 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.193216 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.193235 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.193282 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:41Z","lastTransitionTime":"2025-11-24T08:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.276449 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.276502 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.276516 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.276536 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:41 crc kubenswrapper[4944]: E1124 08:53:41.276610 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:41 crc kubenswrapper[4944]: E1124 08:53:41.276711 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:41 crc kubenswrapper[4944]: E1124 08:53:41.276772 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:41 crc kubenswrapper[4944]: E1124 08:53:41.276843 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.296751 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.296802 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.296812 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.296835 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.296855 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:41Z","lastTransitionTime":"2025-11-24T08:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.399793 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.399848 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.399856 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.399873 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.399883 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:41Z","lastTransitionTime":"2025-11-24T08:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.502598 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.502649 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.502662 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.502684 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.502698 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:41Z","lastTransitionTime":"2025-11-24T08:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.605718 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.605766 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.605776 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.605792 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.605802 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:41Z","lastTransitionTime":"2025-11-24T08:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.708382 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.708435 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.708448 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.708463 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.708474 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:41Z","lastTransitionTime":"2025-11-24T08:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.770019 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/3.log" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.770591 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/2.log" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.773021 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3" exitCode=1 Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.773084 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3"} Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.773131 4944 scope.go:117] "RemoveContainer" containerID="ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.773849 4944 scope.go:117] "RemoveContainer" containerID="e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3" Nov 24 08:53:41 crc kubenswrapper[4944]: E1124 08:53:41.774046 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.792968 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.804011 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.810963 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.811002 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.811017 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.811086 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.811102 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:41Z","lastTransitionTime":"2025-11-24T08:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.815084 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.826252 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bebeee-67e1-41eb-902f-863da7195b46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://739cd0cec0794b22aa306490236c6e42aef2aab4b8787cfed8752e727d1c48c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10a53461679aed9575532a08cb82b6bae353bf170c7a663369aeb4dac512e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94d33673d9208d6570c2853f6644328f2997815a8bfd48a5d7338d12bab2f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.838857 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.850465 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.862004 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.875336 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.889722 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.899843 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.910302 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.912978 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.913016 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.913028 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.913061 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.913076 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:41Z","lastTransitionTime":"2025-11-24T08:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.922839 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.937588 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.950639 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:28Z\\\",\\\"message\\\":\\\"2025-11-24T08:52:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5\\\\n2025-11-24T08:52:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5 to /host/opt/cni/bin/\\\\n2025-11-24T08:52:43Z [verbose] multus-daemon started\\\\n2025-11-24T08:52:43Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:53:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:53:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.967400 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba730ea47730d6d720f54b0229703090e159535dcd2f8127a5693be712046ed1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:12Z\\\",\\\"message\\\":\\\"ches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager/controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager/controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.149\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1124 08:53:12.129940 6626 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:41Z\\\",\\\"message\\\":\\\"41.078677 6996 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:53:41.078982 6996 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:53:41.081235 6996 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:53:41.083805 6996 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:53:41.083835 6996 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 08:53:41.083843 6996 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 08:53:41.083844 6996 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:53:41.083859 6996 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 08:53:41.083867 6996 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 08:53:41.083878 6996 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 08:53:41.084000 6996 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 08:53:41.084036 6996 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:53:41.084081 6996 factory.go:656] Stopping watch factory\\\\nI1124 08:53:41.084094 6996 ovnkube.go:599] Stopped ovnkube\\\\nI1124 08:53:41.084115 6996 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:53:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.977621 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.987499 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:41 crc kubenswrapper[4944]: I1124 08:53:41.997634 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6043c3-7c9c-4773-b43a-51026a0d0ff5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae6958eb78d429d6c25449d6b5595b7ef19937a8db223b185317ae344b8dd947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:41Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.016206 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.016297 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.016315 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.016341 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.016357 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:42Z","lastTransitionTime":"2025-11-24T08:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.118636 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.118682 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.118693 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.118710 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.118722 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:42Z","lastTransitionTime":"2025-11-24T08:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.221402 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.221452 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.221467 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.221499 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.221511 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:42Z","lastTransitionTime":"2025-11-24T08:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.324425 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.324491 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.324506 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.324530 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.324546 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:42Z","lastTransitionTime":"2025-11-24T08:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.427751 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.427793 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.427804 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.427820 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.427833 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:42Z","lastTransitionTime":"2025-11-24T08:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.530959 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.531010 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.531023 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.531077 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.531096 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:42Z","lastTransitionTime":"2025-11-24T08:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.633787 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.633845 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.633860 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.633877 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.633888 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:42Z","lastTransitionTime":"2025-11-24T08:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.742234 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.742311 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.742325 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.742347 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.742363 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:42Z","lastTransitionTime":"2025-11-24T08:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.779704 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/3.log" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.784589 4944 scope.go:117] "RemoveContainer" containerID="e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3" Nov 24 08:53:42 crc kubenswrapper[4944]: E1124 08:53:42.784787 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.803046 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.820101 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.834401 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.846587 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.846666 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.846685 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.846712 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.846731 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:42Z","lastTransitionTime":"2025-11-24T08:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.847606 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.860817 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.872555 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.882439 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.892435 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.902921 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6043c3-7c9c-4773-b43a-51026a0d0ff5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae6958eb78d429d6c25449d6b5595b7ef19937a8db223b185317ae344b8dd947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.914674 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.927827 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:28Z\\\",\\\"message\\\":\\\"2025-11-24T08:52:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5\\\\n2025-11-24T08:52:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5 to /host/opt/cni/bin/\\\\n2025-11-24T08:52:43Z [verbose] multus-daemon started\\\\n2025-11-24T08:52:43Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:53:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:53:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.946969 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:41Z\\\",\\\"message\\\":\\\"41.078677 6996 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:53:41.078982 6996 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:53:41.081235 6996 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:53:41.083805 6996 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:53:41.083835 6996 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 08:53:41.083843 6996 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 08:53:41.083844 6996 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:53:41.083859 6996 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 08:53:41.083867 6996 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 08:53:41.083878 6996 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 08:53:41.084000 6996 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 08:53:41.084036 6996 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:53:41.084081 6996 factory.go:656] Stopping watch factory\\\\nI1124 08:53:41.084094 6996 ovnkube.go:599] Stopped ovnkube\\\\nI1124 08:53:41.084115 6996 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:53:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.949547 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.949622 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.949640 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.949659 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.949671 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:42Z","lastTransitionTime":"2025-11-24T08:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.959669 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.970487 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.982557 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bebeee-67e1-41eb-902f-863da7195b46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://739cd0cec0794b22aa306490236c6e42aef2aab4b8787cfed8752e727d1c48c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10a53461679aed9575532a08cb82b6bae353bf170c7a663369aeb4dac512e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94d33673d9208d6570c2853f6644328f2997815a8bfd48a5d7338d12bab2f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:42 crc kubenswrapper[4944]: I1124 08:53:42.995833 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:42Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.007782 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.018519 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.052787 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.052842 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.052860 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.052882 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.052896 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:43Z","lastTransitionTime":"2025-11-24T08:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.107936 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.108068 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.108116 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.108219 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.108191275 +0000 UTC m=+147.642631737 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.108225 4944 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.108281 4944 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.108295 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.108289368 +0000 UTC m=+147.642729830 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.108351 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.108337539 +0000 UTC m=+147.642778201 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.155004 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.155073 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.155086 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.155104 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.155115 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:43Z","lastTransitionTime":"2025-11-24T08:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.208959 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.209034 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.209428 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.209448 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.209459 4944 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.209460 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.209503 4944 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.209523 4944 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.209527 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.209513 +0000 UTC m=+147.743953462 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.209599 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.209573502 +0000 UTC m=+147.744014144 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.257955 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.257997 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.258006 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.258020 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.258031 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:43Z","lastTransitionTime":"2025-11-24T08:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.275824 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.275824 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.275862 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.275957 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.276112 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.276275 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.276357 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:43 crc kubenswrapper[4944]: E1124 08:53:43.276617 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.360968 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.361013 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.361024 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.361042 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.361069 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:43Z","lastTransitionTime":"2025-11-24T08:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.463339 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.463382 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.463416 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.463432 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.463443 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:43Z","lastTransitionTime":"2025-11-24T08:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.565939 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.565995 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.566012 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.566030 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.566058 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:43Z","lastTransitionTime":"2025-11-24T08:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.668713 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.668768 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.668778 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.668797 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.668808 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:43Z","lastTransitionTime":"2025-11-24T08:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.771290 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.771339 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.771348 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.771363 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.771375 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:43Z","lastTransitionTime":"2025-11-24T08:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.874130 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.874176 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.874186 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.874200 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.874210 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:43Z","lastTransitionTime":"2025-11-24T08:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.976882 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.976931 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.976940 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.976957 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:43 crc kubenswrapper[4944]: I1124 08:53:43.976968 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:43Z","lastTransitionTime":"2025-11-24T08:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.080092 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.080163 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.080184 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.080208 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.080223 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:44Z","lastTransitionTime":"2025-11-24T08:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.183090 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.183153 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.183166 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.183182 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.183194 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:44Z","lastTransitionTime":"2025-11-24T08:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.287524 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.287606 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.287629 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.287656 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.287675 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:44Z","lastTransitionTime":"2025-11-24T08:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.390001 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.390446 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.390455 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.390470 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.390482 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:44Z","lastTransitionTime":"2025-11-24T08:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.492654 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.492702 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.492719 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.492737 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.492748 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:44Z","lastTransitionTime":"2025-11-24T08:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.595782 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.595820 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.595831 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.595848 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.595859 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:44Z","lastTransitionTime":"2025-11-24T08:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.698619 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.698662 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.698672 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.698690 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.698701 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:44Z","lastTransitionTime":"2025-11-24T08:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.801494 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.801537 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.801547 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.801564 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.801574 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:44Z","lastTransitionTime":"2025-11-24T08:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.903476 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.903525 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.903534 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.903552 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:44 crc kubenswrapper[4944]: I1124 08:53:44.903564 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:44Z","lastTransitionTime":"2025-11-24T08:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.006084 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.006166 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.006175 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.006194 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.006207 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:45Z","lastTransitionTime":"2025-11-24T08:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.109087 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.109135 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.109146 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.109163 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.109174 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:45Z","lastTransitionTime":"2025-11-24T08:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.211997 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.212040 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.212079 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.212099 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.212111 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:45Z","lastTransitionTime":"2025-11-24T08:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.276215 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.276279 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.276252 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:45 crc kubenswrapper[4944]: E1124 08:53:45.276405 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.276272 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:45 crc kubenswrapper[4944]: E1124 08:53:45.276486 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:45 crc kubenswrapper[4944]: E1124 08:53:45.276525 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:45 crc kubenswrapper[4944]: E1124 08:53:45.276571 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.314432 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.314477 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.314492 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.314512 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.314525 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:45Z","lastTransitionTime":"2025-11-24T08:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.417027 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.417091 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.417103 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.417121 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.417133 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:45Z","lastTransitionTime":"2025-11-24T08:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.519656 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.519696 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.519708 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.519726 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.519737 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:45Z","lastTransitionTime":"2025-11-24T08:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.622703 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.622752 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.622764 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.622781 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.622793 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:45Z","lastTransitionTime":"2025-11-24T08:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.725255 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.725296 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.725313 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.725330 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.725342 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:45Z","lastTransitionTime":"2025-11-24T08:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.827766 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.827818 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.827839 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.827858 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.827867 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:45Z","lastTransitionTime":"2025-11-24T08:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.930284 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.930355 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.930373 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.930392 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:45 crc kubenswrapper[4944]: I1124 08:53:45.930405 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:45Z","lastTransitionTime":"2025-11-24T08:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.033708 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.033779 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.033791 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.033811 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.033823 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:46Z","lastTransitionTime":"2025-11-24T08:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.136341 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.136382 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.136391 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.136407 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.136419 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:46Z","lastTransitionTime":"2025-11-24T08:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.239291 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.239366 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.239379 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.239400 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.239414 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:46Z","lastTransitionTime":"2025-11-24T08:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.342528 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.343173 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.343230 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.343251 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.343262 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:46Z","lastTransitionTime":"2025-11-24T08:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.445335 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.445742 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.445803 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.445870 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.445934 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:46Z","lastTransitionTime":"2025-11-24T08:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.548653 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.548705 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.548715 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.548735 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.548749 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:46Z","lastTransitionTime":"2025-11-24T08:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.651332 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.651377 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.651388 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.651408 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.651420 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:46Z","lastTransitionTime":"2025-11-24T08:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.754362 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.754410 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.754421 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.754440 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.754452 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:46Z","lastTransitionTime":"2025-11-24T08:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.857293 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.857331 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.857341 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.857358 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.857369 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:46Z","lastTransitionTime":"2025-11-24T08:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.960092 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.960152 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.960163 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.960183 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:46 crc kubenswrapper[4944]: I1124 08:53:46.960201 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:46Z","lastTransitionTime":"2025-11-24T08:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.063523 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.063562 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.063574 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.063594 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.063605 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.165770 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.165817 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.165829 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.165844 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.165855 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.268286 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.268337 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.268351 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.268371 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.268384 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.275783 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.275800 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.275792 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.275783 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:47 crc kubenswrapper[4944]: E1124 08:53:47.275895 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:47 crc kubenswrapper[4944]: E1124 08:53:47.275963 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:47 crc kubenswrapper[4944]: E1124 08:53:47.276057 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:47 crc kubenswrapper[4944]: E1124 08:53:47.276158 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.308040 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.308468 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.308565 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.308736 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.308865 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: E1124 08:53:47.322786 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.326598 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.326633 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.326645 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.326663 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.326673 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: E1124 08:53:47.337256 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.340938 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.340971 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.340982 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.340996 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.341006 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: E1124 08:53:47.350971 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.355562 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.355606 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.355616 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.355634 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.355645 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: E1124 08:53:47.366885 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.370086 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.370195 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.370291 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.370379 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.370462 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: E1124 08:53:47.382380 4944 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2763b4ac-9259-4476-9352-1d70e540eb24\\\",\\\"systemUUID\\\":\\\"095e1c26-41b9-4bd1-9ffb-7095139764e0\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:47 crc kubenswrapper[4944]: E1124 08:53:47.382687 4944 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.384173 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.384255 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.384266 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.384280 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.384289 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.486773 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.487141 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.487214 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.487369 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.487437 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.590549 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.590955 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.591084 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.591184 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.591269 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.694489 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.694553 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.694566 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.694585 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.694599 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.797199 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.797236 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.797247 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.797264 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.797275 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.900194 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.900228 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.900237 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.900251 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:47 crc kubenswrapper[4944]: I1124 08:53:47.900261 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:47Z","lastTransitionTime":"2025-11-24T08:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.002681 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.002717 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.002725 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.002740 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.002751 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:48Z","lastTransitionTime":"2025-11-24T08:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.105534 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.105584 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.105598 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.105617 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.105632 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:48Z","lastTransitionTime":"2025-11-24T08:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.208418 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.208470 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.208487 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.208507 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.208522 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:48Z","lastTransitionTime":"2025-11-24T08:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.311814 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.311862 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.311873 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.311890 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.311903 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:48Z","lastTransitionTime":"2025-11-24T08:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.414867 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.414911 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.414921 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.414938 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.414950 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:48Z","lastTransitionTime":"2025-11-24T08:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.518004 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.518077 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.518093 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.518115 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.518133 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:48Z","lastTransitionTime":"2025-11-24T08:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.621366 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.621409 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.621421 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.621439 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.621449 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:48Z","lastTransitionTime":"2025-11-24T08:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.729537 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.729726 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.729742 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.729764 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.729779 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:48Z","lastTransitionTime":"2025-11-24T08:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.833122 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.833161 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.833179 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.833195 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.833204 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:48Z","lastTransitionTime":"2025-11-24T08:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.938109 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.938170 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.938182 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.938202 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:48 crc kubenswrapper[4944]: I1124 08:53:48.938213 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:48Z","lastTransitionTime":"2025-11-24T08:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.040671 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.040714 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.040725 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.040744 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.040757 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:49Z","lastTransitionTime":"2025-11-24T08:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.143141 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.143182 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.143191 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.143208 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.143218 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:49Z","lastTransitionTime":"2025-11-24T08:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.245265 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.245323 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.245333 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.245350 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.245363 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:49Z","lastTransitionTime":"2025-11-24T08:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.276573 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.276631 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.276611 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.276608 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:49 crc kubenswrapper[4944]: E1124 08:53:49.276758 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:49 crc kubenswrapper[4944]: E1124 08:53:49.276901 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:49 crc kubenswrapper[4944]: E1124 08:53:49.277080 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:49 crc kubenswrapper[4944]: E1124 08:53:49.277119 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.348132 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.348201 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.348211 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.348226 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.348236 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:49Z","lastTransitionTime":"2025-11-24T08:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.451016 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.451084 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.451096 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.451136 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.451149 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:49Z","lastTransitionTime":"2025-11-24T08:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.553694 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.553738 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.553749 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.553771 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.553786 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:49Z","lastTransitionTime":"2025-11-24T08:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.656767 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.656798 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.656806 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.656821 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.656831 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:49Z","lastTransitionTime":"2025-11-24T08:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.759025 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.759084 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.759095 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.759113 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.759126 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:49Z","lastTransitionTime":"2025-11-24T08:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.861794 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.861837 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.861846 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.861863 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.861872 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:49Z","lastTransitionTime":"2025-11-24T08:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.963812 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.963856 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.963869 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.963885 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:49 crc kubenswrapper[4944]: I1124 08:53:49.963896 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:49Z","lastTransitionTime":"2025-11-24T08:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.066710 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.066772 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.066786 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.066809 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.066834 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:50Z","lastTransitionTime":"2025-11-24T08:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.169552 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.169600 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.169611 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.169630 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.169642 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:50Z","lastTransitionTime":"2025-11-24T08:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.271899 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.271961 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.271972 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.271990 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.271999 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:50Z","lastTransitionTime":"2025-11-24T08:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.294977 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d12ca80b-ccaf-42e6-96c5-8476c8b02362\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b32f277138fab4edf143f74879a2a5b49a51f746497e377b2457f81030af3609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65449029f89119dff47b13c304ad1173ea0c0b4397ec72573aa341414d39d3a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6b395a80704cb16da603dc1239ef8b9aa3f7232d4f5cf7e5250f63f99ae61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72bfb3d0f211b5c67a0e67685023718098612e9fb7b878a19ee0468d03691afd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.304994 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://302bd8c2aceec003d0b515a19395529439bb73c09663648fac13cab4d046c53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.314607 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e5c60c-e37d-43c8-860e-39ef930adc16\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69e196f478d7933322ee4f180eff8e5b2a039d4249887565e8bd9b757c785531\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtq7z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-llhq9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.325110 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcff1fc-7f00-49e6-8095-fde5f3c24a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de9e753362a11e6b206f33025198a72bc6401688db511fa1d2ff9972979fe8ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ebbf86074e2afeb2d82f73b50e1c72b7640e75b69a4a4a85ab55ce63a56b374\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f8bdk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-llb5d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.333743 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac6043c3-7c9c-4773-b43a-51026a0d0ff5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae6958eb78d429d6c25449d6b5595b7ef19937a8db223b185317ae344b8dd947\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072b89e9a4d08a535263d58ad66356f74b4fbf3bd3bc223d48150ecb6536a364\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.343994 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.355316 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dg4dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5ff38b2b-032d-47df-8836-105e0cfae835\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:28Z\\\",\\\"message\\\":\\\"2025-11-24T08:52:42+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5\\\\n2025-11-24T08:52:42+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0d4b72d0-5929-4706-b4cb-5258077d71c5 to /host/opt/cni/bin/\\\\n2025-11-24T08:52:43Z [verbose] multus-daemon started\\\\n2025-11-24T08:52:43Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:53:28Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:53:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p26h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dg4dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.373901 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.373965 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.373977 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.373995 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.374007 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:50Z","lastTransitionTime":"2025-11-24T08:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.376063 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb94161-d41a-420b-b5ed-9b65abbe91a3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:53:41Z\\\",\\\"message\\\":\\\"41.078677 6996 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:53:41.078982 6996 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:53:41.081235 6996 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:53:41.083805 6996 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:53:41.083835 6996 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1124 08:53:41.083843 6996 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1124 08:53:41.083844 6996 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:53:41.083859 6996 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1124 08:53:41.083867 6996 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1124 08:53:41.083878 6996 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 08:53:41.084000 6996 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 08:53:41.084036 6996 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:53:41.084081 6996 factory.go:656] Stopping watch factory\\\\nI1124 08:53:41.084094 6996 ovnkube.go:599] Stopped ovnkube\\\\nI1124 08:53:41.084115 6996 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:53:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:53:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn57m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7q5tz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.386635 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2267c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2fbc441-86ee-4abe-bfc9-75f5b8cbb04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f83737364e8005708435aaed2969bb1c2d5cd0d92f570fd84843fe64932b2f23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b9mfc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2267c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.398354 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58bebeee-67e1-41eb-902f-863da7195b46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:53:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://739cd0cec0794b22aa306490236c6e42aef2aab4b8787cfed8752e727d1c48c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10a53461679aed9575532a08cb82b6bae353bf170c7a663369aeb4dac512e06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94d33673d9208d6570c2853f6644328f2997815a8bfd48a5d7338d12bab2f77d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52c2c5dd950b7f78181e219addcc74a2ea161e55b5f608d9020978c4a220578b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.412553 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.422023 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9zglk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c792a6e-0110-4e35-923e-c1a38f32e8f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de93bff049dad4dc11c10fcdc7713e7827813ea94ce0e05d455e569f7c355073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l95s7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:39Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9zglk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.430542 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55990e0c-a6ff-4b18-8b8d-f27542761408\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fsbbf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jmkb9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.446869 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9e367e9-1de1-43d5-9d2b-bc167c9e484f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4119fdc449d0fe717480a1f17d6b571485ad66971ca9a4b2b453f153e8448a38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285c46eb248f33c018b29d08a9e007c0cc927dcd9c6e4d44f39e568cc94919a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90a006e0f838f95dc2f5f1278367034af6a99aa800e9b4ac9124037405ff3611\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://75583c7913854813fc2316dc08633572b5618e391a80223e6a769d02663a94d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f52973288caafa6693085eb0a3a1ff8d827c14867b52d4c41c113a0d5d80fc64\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:52:23.931597 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:52:23.934939 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2051396207/tls.crt::/tmp/serving-cert-2051396207/tls.key\\\\\\\"\\\\nI1124 08:52:39.037274 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:52:39.040377 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:52:39.040403 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:52:39.040431 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:52:39.040437 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:52:39.048814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1124 08:52:39.048839 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1124 08:52:39.048859 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:52:39.048876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:52:39.048880 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:52:39.048884 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:52:39.048887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1124 08:52:39.053192 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f8b481f058aaf3b22a03148d6fab36923f196df6e45ac4967b6be9832076ceb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:23Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31e39b21f2f7d5f2b58dc2dea1bcd8f07dabd5f884667782d0b7b53a8d25f920\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:20Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.458215 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://514da0648a4b67e61d6d7f97f74ab22482cdeea61713bd09a5e652f3815d6e81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.470995 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.478490 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.478537 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.478594 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.478613 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.478645 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:50Z","lastTransitionTime":"2025-11-24T08:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.483683 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6af633f89992d5cdea6c024124fb0e9831ac08ea4dcd827fc41c18d20666ff75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78cc93f8d7ea5f102efc5e9f857936caf50e1590ccc86818bcf3f52a232c369e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.499114 4944 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5899104-758a-40da-9612-7334e6616928\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:52:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b92eb7d21ee0610de228f63d32d0e070b826fc44bedc937a2416dcf9a8067e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:52:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dccef97937e749da55bd427f2862792b6603ca4e09a3b0f8e4f516d999ed27e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1adefa989a8da132d9f233e655ba7728f2bd706d3721a611a6d70e23e028927e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e70188a1e7038efd06c27fdfb40ee97082957d1c0d87740ae6acd6e7757944e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a32f5844ad0ffb76e8bd40eb32533ec0fb5c6d74d661bd9246e078c5e24341d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9a6221017f175b085bdafe4242bf46fda2ec165011385902dd36bc7642ee07a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c51e322f025a5387ef9c5f4ba81d8aa91011108e9cadd6d40d7a8cce53c727c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:52:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bzdlj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:52:40Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-tc6n6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:53:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.581701 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.581792 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.581829 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.581864 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.581886 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:50Z","lastTransitionTime":"2025-11-24T08:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.684609 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.684674 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.684688 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.684706 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.684721 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:50Z","lastTransitionTime":"2025-11-24T08:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.788107 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.788144 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.788159 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.788180 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.788193 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:50Z","lastTransitionTime":"2025-11-24T08:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.890621 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.890662 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.890671 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.890687 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.890699 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:50Z","lastTransitionTime":"2025-11-24T08:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.992767 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.992810 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.992822 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.992837 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:50 crc kubenswrapper[4944]: I1124 08:53:50.992847 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:50Z","lastTransitionTime":"2025-11-24T08:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.095857 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.095906 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.095917 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.095937 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.095949 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:51Z","lastTransitionTime":"2025-11-24T08:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.199144 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.199212 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.199230 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.199257 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.199276 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:51Z","lastTransitionTime":"2025-11-24T08:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.276782 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.276893 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:51 crc kubenswrapper[4944]: E1124 08:53:51.277186 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.277281 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:51 crc kubenswrapper[4944]: E1124 08:53:51.277530 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.277585 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:51 crc kubenswrapper[4944]: E1124 08:53:51.277758 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:51 crc kubenswrapper[4944]: E1124 08:53:51.277803 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.302434 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.302498 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.302517 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.302543 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.302566 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:51Z","lastTransitionTime":"2025-11-24T08:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.405899 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.405956 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.405971 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.405995 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.406012 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:51Z","lastTransitionTime":"2025-11-24T08:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.509160 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.509210 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.509220 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.509240 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.509251 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:51Z","lastTransitionTime":"2025-11-24T08:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.612420 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.612461 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.612470 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.612484 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.612495 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:51Z","lastTransitionTime":"2025-11-24T08:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.715719 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.715761 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.715772 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.715796 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.715807 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:51Z","lastTransitionTime":"2025-11-24T08:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.818375 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.818429 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.818441 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.818461 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.818474 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:51Z","lastTransitionTime":"2025-11-24T08:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.921427 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.921488 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.921501 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.921522 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:51 crc kubenswrapper[4944]: I1124 08:53:51.921543 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:51Z","lastTransitionTime":"2025-11-24T08:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.024179 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.024229 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.024243 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.024263 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.024277 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:52Z","lastTransitionTime":"2025-11-24T08:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.127212 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.127335 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.127350 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.127370 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.127387 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:52Z","lastTransitionTime":"2025-11-24T08:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.229879 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.229931 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.229947 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.229967 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.229982 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:52Z","lastTransitionTime":"2025-11-24T08:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.333443 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.333859 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.333948 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.334036 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.334132 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:52Z","lastTransitionTime":"2025-11-24T08:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.437037 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.437097 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.437108 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.437125 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.437136 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:52Z","lastTransitionTime":"2025-11-24T08:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.539695 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.539739 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.539747 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.539763 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.539773 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:52Z","lastTransitionTime":"2025-11-24T08:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.642795 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.643270 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.643378 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.643491 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.643617 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:52Z","lastTransitionTime":"2025-11-24T08:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.747189 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.747605 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.747733 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.747888 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.747998 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:52Z","lastTransitionTime":"2025-11-24T08:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.851682 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.852132 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.852264 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.852363 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.852443 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:52Z","lastTransitionTime":"2025-11-24T08:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.955400 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.955462 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.955474 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.955492 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:52 crc kubenswrapper[4944]: I1124 08:53:52.955504 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:52Z","lastTransitionTime":"2025-11-24T08:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.058267 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.058311 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.058323 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.058341 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.058385 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:53Z","lastTransitionTime":"2025-11-24T08:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.161089 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.161149 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.161164 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.161186 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.161201 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:53Z","lastTransitionTime":"2025-11-24T08:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.264198 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.264620 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.264736 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.264816 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.264895 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:53Z","lastTransitionTime":"2025-11-24T08:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.275792 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:53 crc kubenswrapper[4944]: E1124 08:53:53.276101 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.275829 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:53 crc kubenswrapper[4944]: E1124 08:53:53.276298 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.275793 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:53 crc kubenswrapper[4944]: E1124 08:53:53.276471 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.275899 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:53 crc kubenswrapper[4944]: E1124 08:53:53.276711 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.368917 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.369382 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.369485 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.369585 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.369686 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:53Z","lastTransitionTime":"2025-11-24T08:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.471756 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.471795 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.471805 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.471821 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.471830 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:53Z","lastTransitionTime":"2025-11-24T08:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.574946 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.575928 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.576190 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.576638 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.576910 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:53Z","lastTransitionTime":"2025-11-24T08:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.679564 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.679614 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.679626 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.679642 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.679652 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:53Z","lastTransitionTime":"2025-11-24T08:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.782743 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.782784 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.782795 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.782814 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.782825 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:53Z","lastTransitionTime":"2025-11-24T08:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.884799 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.884838 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.884846 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.884862 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.884875 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:53Z","lastTransitionTime":"2025-11-24T08:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.987725 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.987786 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.987799 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.987821 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:53 crc kubenswrapper[4944]: I1124 08:53:53.987842 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:53Z","lastTransitionTime":"2025-11-24T08:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.091221 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.091269 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.091279 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.091305 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.091319 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:54Z","lastTransitionTime":"2025-11-24T08:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.194439 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.194497 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.194512 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.194534 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.194550 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:54Z","lastTransitionTime":"2025-11-24T08:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.298235 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.298294 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.298311 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.298338 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.298352 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:54Z","lastTransitionTime":"2025-11-24T08:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.401031 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.401392 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.401702 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.402026 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.402331 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:54Z","lastTransitionTime":"2025-11-24T08:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.506270 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.506602 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.506707 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.506799 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.506888 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:54Z","lastTransitionTime":"2025-11-24T08:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.609419 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.610160 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.610253 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.610339 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.610411 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:54Z","lastTransitionTime":"2025-11-24T08:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.713999 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.714088 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.714103 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.714150 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.714166 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:54Z","lastTransitionTime":"2025-11-24T08:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.818011 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.818117 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.818142 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.818170 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.818236 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:54Z","lastTransitionTime":"2025-11-24T08:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.921381 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.921432 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.921442 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.921462 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:54 crc kubenswrapper[4944]: I1124 08:53:54.921477 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:54Z","lastTransitionTime":"2025-11-24T08:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.025249 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.025310 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.025323 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.025342 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.025359 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:55Z","lastTransitionTime":"2025-11-24T08:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.128041 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.128113 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.128129 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.128154 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.128168 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:55Z","lastTransitionTime":"2025-11-24T08:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.231692 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.231732 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.231744 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.231761 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.231774 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:55Z","lastTransitionTime":"2025-11-24T08:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.275813 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.275813 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.275814 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.275951 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:55 crc kubenswrapper[4944]: E1124 08:53:55.276132 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:55 crc kubenswrapper[4944]: E1124 08:53:55.276519 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:55 crc kubenswrapper[4944]: E1124 08:53:55.276650 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:55 crc kubenswrapper[4944]: E1124 08:53:55.276828 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.276912 4944 scope.go:117] "RemoveContainer" containerID="e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3" Nov 24 08:53:55 crc kubenswrapper[4944]: E1124 08:53:55.277361 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.335101 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.335157 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.335172 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.335192 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.335208 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:55Z","lastTransitionTime":"2025-11-24T08:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.438148 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.438200 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.438211 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.438227 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.438237 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:55Z","lastTransitionTime":"2025-11-24T08:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.541945 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.541997 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.542007 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.542033 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.542069 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:55Z","lastTransitionTime":"2025-11-24T08:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.644880 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.644936 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.644947 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.644966 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.644979 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:55Z","lastTransitionTime":"2025-11-24T08:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.748178 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.748241 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.748255 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.748274 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.748289 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:55Z","lastTransitionTime":"2025-11-24T08:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.851003 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.851090 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.851113 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.851138 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.851153 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:55Z","lastTransitionTime":"2025-11-24T08:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.953859 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.953939 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.953961 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.953993 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:55 crc kubenswrapper[4944]: I1124 08:53:55.954011 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:55Z","lastTransitionTime":"2025-11-24T08:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.056821 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.056870 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.056881 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.056898 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.056911 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:56Z","lastTransitionTime":"2025-11-24T08:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.159338 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.159717 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.159808 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.159900 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.159990 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:56Z","lastTransitionTime":"2025-11-24T08:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.264091 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.264141 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.264152 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.264170 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.264183 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:56Z","lastTransitionTime":"2025-11-24T08:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.290846 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.366844 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.367258 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.367353 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.367461 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.367541 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:56Z","lastTransitionTime":"2025-11-24T08:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.470654 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.470716 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.470731 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.470756 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.470772 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:56Z","lastTransitionTime":"2025-11-24T08:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.573400 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.573456 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.573468 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.573486 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.573499 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:56Z","lastTransitionTime":"2025-11-24T08:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.676996 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.677122 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.677138 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.677163 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.677176 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:56Z","lastTransitionTime":"2025-11-24T08:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.779351 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.779701 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.779768 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.779852 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.779924 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:56Z","lastTransitionTime":"2025-11-24T08:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.881984 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.882028 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.882039 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.882092 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.882104 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:56Z","lastTransitionTime":"2025-11-24T08:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.985853 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.985940 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.985974 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.986006 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:56 crc kubenswrapper[4944]: I1124 08:53:56.986022 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:56Z","lastTransitionTime":"2025-11-24T08:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.090165 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.090552 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.090872 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.091116 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.091286 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:57Z","lastTransitionTime":"2025-11-24T08:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.194090 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.194161 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.194179 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.194207 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.194225 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:57Z","lastTransitionTime":"2025-11-24T08:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.276792 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.276847 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.276889 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:57 crc kubenswrapper[4944]: E1124 08:53:57.276972 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.277108 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:57 crc kubenswrapper[4944]: E1124 08:53:57.277175 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:57 crc kubenswrapper[4944]: E1124 08:53:57.277442 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:57 crc kubenswrapper[4944]: E1124 08:53:57.277639 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.297501 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.297633 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.297648 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.297667 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.297681 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:57Z","lastTransitionTime":"2025-11-24T08:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.401174 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.401232 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.401243 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.401264 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.401277 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:57Z","lastTransitionTime":"2025-11-24T08:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.504705 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.504763 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.504773 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.504790 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.504802 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:57Z","lastTransitionTime":"2025-11-24T08:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.512195 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.512261 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.512273 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.512295 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.512306 4944 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:53:57Z","lastTransitionTime":"2025-11-24T08:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.552603 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:57 crc kubenswrapper[4944]: E1124 08:53:57.552766 4944 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:53:57 crc kubenswrapper[4944]: E1124 08:53:57.552820 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs podName:55990e0c-a6ff-4b18-8b8d-f27542761408 nodeName:}" failed. No retries permitted until 2025-11-24 08:55:01.552804032 +0000 UTC m=+162.087244494 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs") pod "network-metrics-daemon-jmkb9" (UID: "55990e0c-a6ff-4b18-8b8d-f27542761408") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.573077 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f"] Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.573529 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.575788 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.576396 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.576556 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.576716 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.609712 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2267c" podStartSLOduration=78.609686135 podStartE2EDuration="1m18.609686135s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:57.595601753 +0000 UTC m=+98.130042215" watchObservedRunningTime="2025-11-24 08:53:57.609686135 +0000 UTC m=+98.144126597" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.609932 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-llb5d" podStartSLOduration=78.609924402 podStartE2EDuration="1m18.609924402s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:57.609509729 +0000 UTC m=+98.143950201" watchObservedRunningTime="2025-11-24 08:53:57.609924402 +0000 UTC m=+98.144364864" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.621292 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=28.621265586 podStartE2EDuration="28.621265586s" podCreationTimestamp="2025-11-24 08:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:57.621196954 +0000 UTC m=+98.155637406" watchObservedRunningTime="2025-11-24 08:53:57.621265586 +0000 UTC m=+98.155706048" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.645398 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.645378008 podStartE2EDuration="1.645378008s" podCreationTimestamp="2025-11-24 08:53:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:57.645350287 +0000 UTC m=+98.179790769" watchObservedRunningTime="2025-11-24 08:53:57.645378008 +0000 UTC m=+98.179818470" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.653487 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/edb2649f-805a-4b7d-88f9-654420eefa0a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.653545 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/edb2649f-805a-4b7d-88f9-654420eefa0a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.653740 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/edb2649f-805a-4b7d-88f9-654420eefa0a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.653803 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/edb2649f-805a-4b7d-88f9-654420eefa0a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.653890 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edb2649f-805a-4b7d-88f9-654420eefa0a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.678258 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-dg4dk" podStartSLOduration=78.67823498 podStartE2EDuration="1m18.67823498s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:57.678037354 +0000 UTC m=+98.212477836" watchObservedRunningTime="2025-11-24 08:53:57.67823498 +0000 UTC m=+98.212675442" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.742428 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=41.742407477 podStartE2EDuration="41.742407477s" podCreationTimestamp="2025-11-24 08:53:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:57.742241851 +0000 UTC m=+98.276682333" watchObservedRunningTime="2025-11-24 08:53:57.742407477 +0000 UTC m=+98.276847939" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.755450 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/edb2649f-805a-4b7d-88f9-654420eefa0a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.755505 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/edb2649f-805a-4b7d-88f9-654420eefa0a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.755572 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/edb2649f-805a-4b7d-88f9-654420eefa0a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.755598 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/edb2649f-805a-4b7d-88f9-654420eefa0a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.755638 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edb2649f-805a-4b7d-88f9-654420eefa0a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.755663 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/edb2649f-805a-4b7d-88f9-654420eefa0a-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.755759 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/edb2649f-805a-4b7d-88f9-654420eefa0a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.756823 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/edb2649f-805a-4b7d-88f9-654420eefa0a-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.762556 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edb2649f-805a-4b7d-88f9-654420eefa0a-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.767776 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-9zglk" podStartSLOduration=78.767739148 podStartE2EDuration="1m18.767739148s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:57.766876261 +0000 UTC m=+98.301316743" watchObservedRunningTime="2025-11-24 08:53:57.767739148 +0000 UTC m=+98.302179610" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.780684 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/edb2649f-805a-4b7d-88f9-654420eefa0a-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4qm8f\" (UID: \"edb2649f-805a-4b7d-88f9-654420eefa0a\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.814460 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.814377353 podStartE2EDuration="1m18.814377353s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:57.810972134 +0000 UTC m=+98.345412616" watchObservedRunningTime="2025-11-24 08:53:57.814377353 +0000 UTC m=+98.348817815" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.872529 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-tc6n6" podStartSLOduration=78.872504946 podStartE2EDuration="1m18.872504946s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:57.871662068 +0000 UTC m=+98.406102540" watchObservedRunningTime="2025-11-24 08:53:57.872504946 +0000 UTC m=+98.406945408" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.889687 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.902350 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.902329281 podStartE2EDuration="1m18.902329281s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:57.884880872 +0000 UTC m=+98.419321354" watchObservedRunningTime="2025-11-24 08:53:57.902329281 +0000 UTC m=+98.436769743" Nov 24 08:53:57 crc kubenswrapper[4944]: I1124 08:53:57.922065 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podStartSLOduration=78.922031172 podStartE2EDuration="1m18.922031172s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:57.921391971 +0000 UTC m=+98.455832453" watchObservedRunningTime="2025-11-24 08:53:57.922031172 +0000 UTC m=+98.456471634" Nov 24 08:53:58 crc kubenswrapper[4944]: I1124 08:53:58.843210 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" event={"ID":"edb2649f-805a-4b7d-88f9-654420eefa0a","Type":"ContainerStarted","Data":"70ec8fb9d0a7b3aefb5264f5cf37dc589c50d094f7f1e68d005c84e0150f9f3f"} Nov 24 08:53:58 crc kubenswrapper[4944]: I1124 08:53:58.844463 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" event={"ID":"edb2649f-805a-4b7d-88f9-654420eefa0a","Type":"ContainerStarted","Data":"12b142326bc6797c0156c5bf3880a6630b83e93947c4860508aecc05429bee81"} Nov 24 08:53:58 crc kubenswrapper[4944]: I1124 08:53:58.861399 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4qm8f" podStartSLOduration=79.861370661 podStartE2EDuration="1m19.861370661s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:53:58.861256797 +0000 UTC m=+99.395697279" watchObservedRunningTime="2025-11-24 08:53:58.861370661 +0000 UTC m=+99.395811123" Nov 24 08:53:59 crc kubenswrapper[4944]: I1124 08:53:59.276317 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:53:59 crc kubenswrapper[4944]: I1124 08:53:59.276352 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:53:59 crc kubenswrapper[4944]: I1124 08:53:59.276384 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:53:59 crc kubenswrapper[4944]: I1124 08:53:59.276426 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:53:59 crc kubenswrapper[4944]: E1124 08:53:59.276478 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:53:59 crc kubenswrapper[4944]: E1124 08:53:59.276623 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:53:59 crc kubenswrapper[4944]: E1124 08:53:59.276790 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:53:59 crc kubenswrapper[4944]: E1124 08:53:59.276932 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:01 crc kubenswrapper[4944]: I1124 08:54:01.275742 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:01 crc kubenswrapper[4944]: I1124 08:54:01.275793 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:01 crc kubenswrapper[4944]: I1124 08:54:01.275836 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:01 crc kubenswrapper[4944]: E1124 08:54:01.276432 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:01 crc kubenswrapper[4944]: E1124 08:54:01.276249 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:01 crc kubenswrapper[4944]: I1124 08:54:01.275884 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:01 crc kubenswrapper[4944]: E1124 08:54:01.276505 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:01 crc kubenswrapper[4944]: E1124 08:54:01.276600 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:03 crc kubenswrapper[4944]: I1124 08:54:03.276160 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:03 crc kubenswrapper[4944]: I1124 08:54:03.276293 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:03 crc kubenswrapper[4944]: E1124 08:54:03.276953 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:03 crc kubenswrapper[4944]: I1124 08:54:03.276331 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:03 crc kubenswrapper[4944]: E1124 08:54:03.277025 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:03 crc kubenswrapper[4944]: I1124 08:54:03.276293 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:03 crc kubenswrapper[4944]: E1124 08:54:03.277099 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:03 crc kubenswrapper[4944]: E1124 08:54:03.276876 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:05 crc kubenswrapper[4944]: I1124 08:54:05.276408 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:05 crc kubenswrapper[4944]: I1124 08:54:05.276480 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:05 crc kubenswrapper[4944]: I1124 08:54:05.276535 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:05 crc kubenswrapper[4944]: I1124 08:54:05.276542 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:05 crc kubenswrapper[4944]: E1124 08:54:05.276701 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:05 crc kubenswrapper[4944]: E1124 08:54:05.276817 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:05 crc kubenswrapper[4944]: E1124 08:54:05.276939 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:05 crc kubenswrapper[4944]: E1124 08:54:05.277235 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:07 crc kubenswrapper[4944]: I1124 08:54:07.276830 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:07 crc kubenswrapper[4944]: I1124 08:54:07.276886 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:07 crc kubenswrapper[4944]: I1124 08:54:07.276832 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:07 crc kubenswrapper[4944]: E1124 08:54:07.277030 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:07 crc kubenswrapper[4944]: E1124 08:54:07.277165 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:07 crc kubenswrapper[4944]: E1124 08:54:07.277330 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:07 crc kubenswrapper[4944]: I1124 08:54:07.277691 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:07 crc kubenswrapper[4944]: E1124 08:54:07.277847 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:08 crc kubenswrapper[4944]: I1124 08:54:08.276815 4944 scope.go:117] "RemoveContainer" containerID="e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3" Nov 24 08:54:08 crc kubenswrapper[4944]: E1124 08:54:08.276987 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" Nov 24 08:54:09 crc kubenswrapper[4944]: I1124 08:54:09.276733 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:09 crc kubenswrapper[4944]: I1124 08:54:09.276891 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:09 crc kubenswrapper[4944]: E1124 08:54:09.276913 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:09 crc kubenswrapper[4944]: E1124 08:54:09.277143 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:09 crc kubenswrapper[4944]: I1124 08:54:09.277282 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:09 crc kubenswrapper[4944]: I1124 08:54:09.277627 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:09 crc kubenswrapper[4944]: E1124 08:54:09.277754 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:09 crc kubenswrapper[4944]: E1124 08:54:09.277996 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:11 crc kubenswrapper[4944]: I1124 08:54:11.276476 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:11 crc kubenswrapper[4944]: I1124 08:54:11.276683 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:11 crc kubenswrapper[4944]: E1124 08:54:11.276867 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:11 crc kubenswrapper[4944]: I1124 08:54:11.276960 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:11 crc kubenswrapper[4944]: E1124 08:54:11.277151 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:11 crc kubenswrapper[4944]: I1124 08:54:11.277240 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:11 crc kubenswrapper[4944]: E1124 08:54:11.277277 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:11 crc kubenswrapper[4944]: E1124 08:54:11.277530 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:13 crc kubenswrapper[4944]: I1124 08:54:13.276786 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:13 crc kubenswrapper[4944]: I1124 08:54:13.276839 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:13 crc kubenswrapper[4944]: I1124 08:54:13.276845 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:13 crc kubenswrapper[4944]: I1124 08:54:13.276799 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:13 crc kubenswrapper[4944]: E1124 08:54:13.276952 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:13 crc kubenswrapper[4944]: E1124 08:54:13.277005 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:13 crc kubenswrapper[4944]: E1124 08:54:13.277092 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:13 crc kubenswrapper[4944]: E1124 08:54:13.277161 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:14 crc kubenswrapper[4944]: I1124 08:54:14.897753 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dg4dk_5ff38b2b-032d-47df-8836-105e0cfae835/kube-multus/1.log" Nov 24 08:54:14 crc kubenswrapper[4944]: I1124 08:54:14.898246 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dg4dk_5ff38b2b-032d-47df-8836-105e0cfae835/kube-multus/0.log" Nov 24 08:54:14 crc kubenswrapper[4944]: I1124 08:54:14.898283 4944 generic.go:334] "Generic (PLEG): container finished" podID="5ff38b2b-032d-47df-8836-105e0cfae835" containerID="df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f" exitCode=1 Nov 24 08:54:14 crc kubenswrapper[4944]: I1124 08:54:14.898313 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dg4dk" event={"ID":"5ff38b2b-032d-47df-8836-105e0cfae835","Type":"ContainerDied","Data":"df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f"} Nov 24 08:54:14 crc kubenswrapper[4944]: I1124 08:54:14.898350 4944 scope.go:117] "RemoveContainer" containerID="d93c01ccc8be261b96d0c4662c0e217bfad47bf95b90eeca98a5a0f7920560c6" Nov 24 08:54:14 crc kubenswrapper[4944]: I1124 08:54:14.898774 4944 scope.go:117] "RemoveContainer" containerID="df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f" Nov 24 08:54:14 crc kubenswrapper[4944]: E1124 08:54:14.898945 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-dg4dk_openshift-multus(5ff38b2b-032d-47df-8836-105e0cfae835)\"" pod="openshift-multus/multus-dg4dk" podUID="5ff38b2b-032d-47df-8836-105e0cfae835" Nov 24 08:54:15 crc kubenswrapper[4944]: I1124 08:54:15.276282 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:15 crc kubenswrapper[4944]: E1124 08:54:15.277155 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:15 crc kubenswrapper[4944]: I1124 08:54:15.276303 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:15 crc kubenswrapper[4944]: E1124 08:54:15.277353 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:15 crc kubenswrapper[4944]: I1124 08:54:15.276302 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:15 crc kubenswrapper[4944]: I1124 08:54:15.276280 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:15 crc kubenswrapper[4944]: E1124 08:54:15.277570 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:15 crc kubenswrapper[4944]: E1124 08:54:15.277736 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:15 crc kubenswrapper[4944]: I1124 08:54:15.904999 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dg4dk_5ff38b2b-032d-47df-8836-105e0cfae835/kube-multus/1.log" Nov 24 08:54:17 crc kubenswrapper[4944]: I1124 08:54:17.276642 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:17 crc kubenswrapper[4944]: I1124 08:54:17.276690 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:17 crc kubenswrapper[4944]: I1124 08:54:17.276790 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:17 crc kubenswrapper[4944]: I1124 08:54:17.276643 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:17 crc kubenswrapper[4944]: E1124 08:54:17.276822 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:17 crc kubenswrapper[4944]: E1124 08:54:17.277035 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:17 crc kubenswrapper[4944]: E1124 08:54:17.277133 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:17 crc kubenswrapper[4944]: E1124 08:54:17.277310 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:19 crc kubenswrapper[4944]: I1124 08:54:19.275912 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:19 crc kubenswrapper[4944]: E1124 08:54:19.276073 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:19 crc kubenswrapper[4944]: I1124 08:54:19.276289 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:19 crc kubenswrapper[4944]: I1124 08:54:19.275912 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:19 crc kubenswrapper[4944]: I1124 08:54:19.276871 4944 scope.go:117] "RemoveContainer" containerID="e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3" Nov 24 08:54:19 crc kubenswrapper[4944]: E1124 08:54:19.277028 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7q5tz_openshift-ovn-kubernetes(aeb94161-d41a-420b-b5ed-9b65abbe91a3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" Nov 24 08:54:19 crc kubenswrapper[4944]: I1124 08:54:19.277159 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:19 crc kubenswrapper[4944]: E1124 08:54:19.277235 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:19 crc kubenswrapper[4944]: E1124 08:54:19.277387 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:19 crc kubenswrapper[4944]: E1124 08:54:19.277572 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:20 crc kubenswrapper[4944]: E1124 08:54:20.275202 4944 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 08:54:20 crc kubenswrapper[4944]: E1124 08:54:20.437029 4944 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 08:54:21 crc kubenswrapper[4944]: I1124 08:54:21.276028 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:21 crc kubenswrapper[4944]: I1124 08:54:21.276101 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:21 crc kubenswrapper[4944]: E1124 08:54:21.276314 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:21 crc kubenswrapper[4944]: I1124 08:54:21.276130 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:21 crc kubenswrapper[4944]: E1124 08:54:21.276388 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:21 crc kubenswrapper[4944]: E1124 08:54:21.276450 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:21 crc kubenswrapper[4944]: I1124 08:54:21.276686 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:21 crc kubenswrapper[4944]: E1124 08:54:21.276835 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:23 crc kubenswrapper[4944]: I1124 08:54:23.276735 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:23 crc kubenswrapper[4944]: I1124 08:54:23.276821 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:23 crc kubenswrapper[4944]: I1124 08:54:23.276748 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:23 crc kubenswrapper[4944]: E1124 08:54:23.276902 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:23 crc kubenswrapper[4944]: I1124 08:54:23.276735 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:23 crc kubenswrapper[4944]: E1124 08:54:23.277001 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:23 crc kubenswrapper[4944]: E1124 08:54:23.277110 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:23 crc kubenswrapper[4944]: E1124 08:54:23.277181 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:25 crc kubenswrapper[4944]: I1124 08:54:25.276007 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:25 crc kubenswrapper[4944]: I1124 08:54:25.276133 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:25 crc kubenswrapper[4944]: I1124 08:54:25.276007 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:25 crc kubenswrapper[4944]: E1124 08:54:25.276301 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:25 crc kubenswrapper[4944]: I1124 08:54:25.276028 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:25 crc kubenswrapper[4944]: E1124 08:54:25.276201 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:25 crc kubenswrapper[4944]: E1124 08:54:25.276492 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:25 crc kubenswrapper[4944]: E1124 08:54:25.276459 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:25 crc kubenswrapper[4944]: E1124 08:54:25.438142 4944 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 08:54:27 crc kubenswrapper[4944]: I1124 08:54:27.275861 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:27 crc kubenswrapper[4944]: I1124 08:54:27.275991 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:27 crc kubenswrapper[4944]: E1124 08:54:27.276015 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:27 crc kubenswrapper[4944]: I1124 08:54:27.275862 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:27 crc kubenswrapper[4944]: E1124 08:54:27.276190 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:27 crc kubenswrapper[4944]: I1124 08:54:27.276219 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:27 crc kubenswrapper[4944]: E1124 08:54:27.276399 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:27 crc kubenswrapper[4944]: E1124 08:54:27.276292 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:29 crc kubenswrapper[4944]: I1124 08:54:29.276027 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:29 crc kubenswrapper[4944]: I1124 08:54:29.276184 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:29 crc kubenswrapper[4944]: E1124 08:54:29.276712 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:29 crc kubenswrapper[4944]: I1124 08:54:29.276243 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:29 crc kubenswrapper[4944]: I1124 08:54:29.276452 4944 scope.go:117] "RemoveContainer" containerID="df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f" Nov 24 08:54:29 crc kubenswrapper[4944]: E1124 08:54:29.276800 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:29 crc kubenswrapper[4944]: I1124 08:54:29.276235 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:29 crc kubenswrapper[4944]: E1124 08:54:29.276886 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:29 crc kubenswrapper[4944]: E1124 08:54:29.276948 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:29 crc kubenswrapper[4944]: I1124 08:54:29.948274 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dg4dk_5ff38b2b-032d-47df-8836-105e0cfae835/kube-multus/1.log" Nov 24 08:54:29 crc kubenswrapper[4944]: I1124 08:54:29.949027 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dg4dk" event={"ID":"5ff38b2b-032d-47df-8836-105e0cfae835","Type":"ContainerStarted","Data":"010a18b9e549761ace384c33053c30fa54319e5b84f6acebd70957282468e242"} Nov 24 08:54:30 crc kubenswrapper[4944]: E1124 08:54:30.439870 4944 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 08:54:31 crc kubenswrapper[4944]: I1124 08:54:31.275845 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:31 crc kubenswrapper[4944]: I1124 08:54:31.275901 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:31 crc kubenswrapper[4944]: E1124 08:54:31.276618 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:31 crc kubenswrapper[4944]: I1124 08:54:31.275977 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:31 crc kubenswrapper[4944]: E1124 08:54:31.276709 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:31 crc kubenswrapper[4944]: I1124 08:54:31.275942 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:31 crc kubenswrapper[4944]: E1124 08:54:31.276761 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:31 crc kubenswrapper[4944]: E1124 08:54:31.276621 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:32 crc kubenswrapper[4944]: I1124 08:54:32.277429 4944 scope.go:117] "RemoveContainer" containerID="e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3" Nov 24 08:54:32 crc kubenswrapper[4944]: I1124 08:54:32.963357 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/3.log" Nov 24 08:54:32 crc kubenswrapper[4944]: I1124 08:54:32.966540 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerStarted","Data":"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb"} Nov 24 08:54:32 crc kubenswrapper[4944]: I1124 08:54:32.967123 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:54:32 crc kubenswrapper[4944]: I1124 08:54:32.999079 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podStartSLOduration=113.999059285 podStartE2EDuration="1m53.999059285s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:32.998156848 +0000 UTC m=+133.532597330" watchObservedRunningTime="2025-11-24 08:54:32.999059285 +0000 UTC m=+133.533499747" Nov 24 08:54:33 crc kubenswrapper[4944]: I1124 08:54:33.194871 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jmkb9"] Nov 24 08:54:33 crc kubenswrapper[4944]: I1124 08:54:33.195078 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:33 crc kubenswrapper[4944]: E1124 08:54:33.195239 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:33 crc kubenswrapper[4944]: I1124 08:54:33.276625 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:33 crc kubenswrapper[4944]: I1124 08:54:33.276731 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:33 crc kubenswrapper[4944]: E1124 08:54:33.276762 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:33 crc kubenswrapper[4944]: I1124 08:54:33.276625 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:33 crc kubenswrapper[4944]: E1124 08:54:33.276879 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:33 crc kubenswrapper[4944]: E1124 08:54:33.276947 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:35 crc kubenswrapper[4944]: I1124 08:54:35.276118 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:35 crc kubenswrapper[4944]: I1124 08:54:35.276118 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:35 crc kubenswrapper[4944]: I1124 08:54:35.276118 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:35 crc kubenswrapper[4944]: I1124 08:54:35.276149 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:35 crc kubenswrapper[4944]: E1124 08:54:35.276478 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jmkb9" podUID="55990e0c-a6ff-4b18-8b8d-f27542761408" Nov 24 08:54:35 crc kubenswrapper[4944]: E1124 08:54:35.276612 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:54:35 crc kubenswrapper[4944]: E1124 08:54:35.276673 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:54:35 crc kubenswrapper[4944]: E1124 08:54:35.276783 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:54:37 crc kubenswrapper[4944]: I1124 08:54:37.276545 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:54:37 crc kubenswrapper[4944]: I1124 08:54:37.276568 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:37 crc kubenswrapper[4944]: I1124 08:54:37.276593 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:37 crc kubenswrapper[4944]: I1124 08:54:37.276756 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:37 crc kubenswrapper[4944]: I1124 08:54:37.279378 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 08:54:37 crc kubenswrapper[4944]: I1124 08:54:37.279514 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 08:54:37 crc kubenswrapper[4944]: I1124 08:54:37.279565 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 08:54:37 crc kubenswrapper[4944]: I1124 08:54:37.279641 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 08:54:37 crc kubenswrapper[4944]: I1124 08:54:37.279711 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 08:54:37 crc kubenswrapper[4944]: I1124 08:54:37.281623 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.654583 4944 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.692956 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tm2ss"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.693474 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-h6vpv"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.693515 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.694814 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.695457 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.695779 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.698340 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hf8bj"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.698964 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.699459 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.699904 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.701207 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.701393 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.702151 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.702179 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.702307 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.702162 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.702569 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.702835 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.702973 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.703211 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.703485 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-v2r8w"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.704110 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.704704 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-9tglc"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.705028 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9tglc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.707144 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.707312 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.707343 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.707445 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.707561 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.708068 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.709227 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-gs29j"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.709845 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.710092 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gg465"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.710478 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.714884 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.735000 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.735398 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.737035 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-s9r2p"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.737266 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.737401 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dbf5z"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.737471 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.737661 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.737722 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.737833 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.737905 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.737991 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738125 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738151 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738208 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738296 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738317 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738345 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738457 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738657 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738678 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738797 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738823 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738839 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738840 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738799 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738909 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738940 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738804 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.739012 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.739244 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.738960 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.739964 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vsszz"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.740123 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.740216 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.740317 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.740427 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.740598 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.740979 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741253 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741407 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741542 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741652 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.742691 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741663 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741717 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741774 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741815 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741837 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741898 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741951 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.741977 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.743914 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.742003 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.744184 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-h6vpv"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.744253 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.744277 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.744360 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.749648 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.750883 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.754962 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.755154 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.755269 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.755298 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.755338 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.755299 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.755479 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.755544 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.755686 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.756453 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.758780 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.759365 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.759584 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.759588 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.761802 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.762139 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.763311 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.763640 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.763958 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.765166 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.765431 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.765563 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.774470 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.774689 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.778221 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.782903 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.783432 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v5288"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.783967 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.783474 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.784257 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.785502 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.802839 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.803954 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.804279 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.805727 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.806199 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-bg5vl"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.806675 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.807221 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.807429 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.807578 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.808615 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.809387 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.809511 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.809591 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810033 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgwdn\" (UniqueName: \"kubernetes.io/projected/e6d222cd-e154-421b-afb4-dbad33103c9a-kube-api-access-dgwdn\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810089 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb59aca8-007d-46f0-a410-75fbe41c7217-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810122 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-dir\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810143 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810165 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810187 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp7xs\" (UniqueName: \"kubernetes.io/projected/b355df00-f706-42c5-b16f-8c237a5048d8-kube-api-access-kp7xs\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810209 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/293773da-e85d-43bc-b06a-19977c55c25a-config\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810223 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g79vv\" (UniqueName: \"kubernetes.io/projected/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-kube-api-access-g79vv\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810241 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810256 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-config\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810271 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bb59aca8-007d-46f0-a410-75fbe41c7217-audit-policies\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810287 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e48ec747-cb70-428b-9dfe-711d563813a0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-szg4v\" (UID: \"e48ec747-cb70-428b-9dfe-711d563813a0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810302 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-etcd-ca\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810317 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810334 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-config\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810352 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sgvz\" (UniqueName: \"kubernetes.io/projected/e33b0db2-705e-4424-985c-61d443a4cded-kube-api-access-7sgvz\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810372 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6c75ce80-6a91-407b-8035-97bc3149d046-machine-approver-tls\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810391 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb59aca8-007d-46f0-a410-75fbe41c7217-serving-cert\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810410 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8123393-4f9a-4546-8c76-10c87169eab3-service-ca-bundle\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810428 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms79n\" (UniqueName: \"kubernetes.io/projected/12954baa-8292-40c1-982a-28e23800078e-kube-api-access-ms79n\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810444 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/201c725b-a236-416d-8d93-2d0b21a103e0-metrics-tls\") pod \"dns-operator-744455d44c-hf8bj\" (UID: \"201c725b-a236-416d-8d93-2d0b21a103e0\") " pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810468 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810484 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6d222cd-e154-421b-afb4-dbad33103c9a-config\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810502 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810519 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8123393-4f9a-4546-8c76-10c87169eab3-serving-cert\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810534 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c75ce80-6a91-407b-8035-97bc3149d046-auth-proxy-config\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810548 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whtlb\" (UniqueName: \"kubernetes.io/projected/6c75ce80-6a91-407b-8035-97bc3149d046-kube-api-access-whtlb\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810563 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c75ce80-6a91-407b-8035-97bc3149d046-config\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810577 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/293773da-e85d-43bc-b06a-19977c55c25a-serving-cert\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810592 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67n5z\" (UniqueName: \"kubernetes.io/projected/293773da-e85d-43bc-b06a-19977c55c25a-kube-api-access-67n5z\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810604 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-etcd-service-ca\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810619 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e6d222cd-e154-421b-afb4-dbad33103c9a-images\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810637 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810653 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12954baa-8292-40c1-982a-28e23800078e-serving-cert\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810670 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810684 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bb59aca8-007d-46f0-a410-75fbe41c7217-audit-dir\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810701 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ms8j\" (UniqueName: \"kubernetes.io/projected/aeedd47e-ece2-4717-839b-755f1e1cd8e0-kube-api-access-4ms8j\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810717 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spnpz\" (UniqueName: \"kubernetes.io/projected/201c725b-a236-416d-8d93-2d0b21a103e0-kube-api-access-spnpz\") pod \"dns-operator-744455d44c-hf8bj\" (UID: \"201c725b-a236-416d-8d93-2d0b21a103e0\") " pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810732 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-console-config\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810756 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-service-ca\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810771 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb59aca8-007d-46f0-a410-75fbe41c7217-etcd-client\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810785 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-etcd-client\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810799 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-client-ca\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810814 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6d222cd-e154-421b-afb4-dbad33103c9a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810830 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-oauth-serving-cert\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810845 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8123393-4f9a-4546-8c76-10c87169eab3-config\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810860 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-serving-cert\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810875 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e48ec747-cb70-428b-9dfe-711d563813a0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-szg4v\" (UID: \"e48ec747-cb70-428b-9dfe-711d563813a0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810891 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/293773da-e85d-43bc-b06a-19977c55c25a-trusted-ca\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810904 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-serving-cert\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810926 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-config\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810941 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810957 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bb59aca8-007d-46f0-a410-75fbe41c7217-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810973 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8gvh\" (UniqueName: \"kubernetes.io/projected/e48ec747-cb70-428b-9dfe-711d563813a0-kube-api-access-k8gvh\") pod \"openshift-apiserver-operator-796bbdcf4f-szg4v\" (UID: \"e48ec747-cb70-428b-9dfe-711d563813a0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.810996 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-trusted-ca-bundle\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.811010 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bb59aca8-007d-46f0-a410-75fbe41c7217-encryption-config\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.811025 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8123393-4f9a-4546-8c76-10c87169eab3-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.811038 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd2nq\" (UniqueName: \"kubernetes.io/projected/d8123393-4f9a-4546-8c76-10c87169eab3-kube-api-access-sd2nq\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.822599 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-oauth-config\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.822719 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg8nc\" (UniqueName: \"kubernetes.io/projected/bb59aca8-007d-46f0-a410-75fbe41c7217-kube-api-access-gg8nc\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.822771 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e33b0db2-705e-4424-985c-61d443a4cded-serving-cert\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.822794 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.822824 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-policies\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.822839 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.822854 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.822875 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-client-ca\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.813289 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.816144 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.813409 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.823976 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.813610 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.815210 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.824303 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.824607 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tm2ss"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.824731 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.825577 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.826778 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.826987 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.832939 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.833619 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.834820 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fqjd6"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.835677 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.842250 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.844121 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m6g58"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.845091 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.845268 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.846104 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.846423 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.846740 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.847457 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.847990 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.850095 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.850540 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.850630 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.852423 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.853233 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.856634 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.857718 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9tglc"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.858976 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.861358 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.861451 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.861880 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.863724 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.867353 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.867900 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.868076 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.872180 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.873069 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-f9snq"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.874126 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-qwbqn"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.875487 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qwbqn" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.879532 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.880915 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.881041 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.881786 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.880920 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.884432 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.885679 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.896180 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.897836 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.898852 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.904912 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gg465"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.906296 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-v2r8w"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.908884 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hf8bj"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.908957 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-gs29j"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.911539 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.913251 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.914767 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.916652 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dbf5z"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.918490 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.918750 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.920307 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-s9r2p"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.921515 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.923116 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m6g58"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.923549 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924278 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924293 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924328 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp7xs\" (UniqueName: \"kubernetes.io/projected/b355df00-f706-42c5-b16f-8c237a5048d8-kube-api-access-kp7xs\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924358 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/293773da-e85d-43bc-b06a-19977c55c25a-config\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924392 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g79vv\" (UniqueName: \"kubernetes.io/projected/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-kube-api-access-g79vv\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924418 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924445 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-config\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924468 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bb59aca8-007d-46f0-a410-75fbe41c7217-audit-policies\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924491 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e48ec747-cb70-428b-9dfe-711d563813a0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-szg4v\" (UID: \"e48ec747-cb70-428b-9dfe-711d563813a0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924517 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-etcd-ca\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924538 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924559 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-config\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924580 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sgvz\" (UniqueName: \"kubernetes.io/projected/e33b0db2-705e-4424-985c-61d443a4cded-kube-api-access-7sgvz\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924604 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6c75ce80-6a91-407b-8035-97bc3149d046-machine-approver-tls\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924628 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8123393-4f9a-4546-8c76-10c87169eab3-service-ca-bundle\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924648 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb59aca8-007d-46f0-a410-75fbe41c7217-serving-cert\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924668 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms79n\" (UniqueName: \"kubernetes.io/projected/12954baa-8292-40c1-982a-28e23800078e-kube-api-access-ms79n\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924690 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/201c725b-a236-416d-8d93-2d0b21a103e0-metrics-tls\") pod \"dns-operator-744455d44c-hf8bj\" (UID: \"201c725b-a236-416d-8d93-2d0b21a103e0\") " pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924713 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924732 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924750 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6d222cd-e154-421b-afb4-dbad33103c9a-config\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924814 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8123393-4f9a-4546-8c76-10c87169eab3-serving-cert\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924836 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c75ce80-6a91-407b-8035-97bc3149d046-auth-proxy-config\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924855 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whtlb\" (UniqueName: \"kubernetes.io/projected/6c75ce80-6a91-407b-8035-97bc3149d046-kube-api-access-whtlb\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924874 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/293773da-e85d-43bc-b06a-19977c55c25a-serving-cert\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924893 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67n5z\" (UniqueName: \"kubernetes.io/projected/293773da-e85d-43bc-b06a-19977c55c25a-kube-api-access-67n5z\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924910 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-etcd-service-ca\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924931 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c75ce80-6a91-407b-8035-97bc3149d046-config\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924951 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e6d222cd-e154-421b-afb4-dbad33103c9a-images\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924973 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.924992 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12954baa-8292-40c1-982a-28e23800078e-serving-cert\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925016 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925038 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bb59aca8-007d-46f0-a410-75fbe41c7217-audit-dir\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925078 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ms8j\" (UniqueName: \"kubernetes.io/projected/aeedd47e-ece2-4717-839b-755f1e1cd8e0-kube-api-access-4ms8j\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925101 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spnpz\" (UniqueName: \"kubernetes.io/projected/201c725b-a236-416d-8d93-2d0b21a103e0-kube-api-access-spnpz\") pod \"dns-operator-744455d44c-hf8bj\" (UID: \"201c725b-a236-416d-8d93-2d0b21a103e0\") " pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925122 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-console-config\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925157 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-service-ca\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925178 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb59aca8-007d-46f0-a410-75fbe41c7217-etcd-client\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925197 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-etcd-client\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925216 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-client-ca\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925235 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6d222cd-e154-421b-afb4-dbad33103c9a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925791 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-oauth-serving-cert\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925825 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e48ec747-cb70-428b-9dfe-711d563813a0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-szg4v\" (UID: \"e48ec747-cb70-428b-9dfe-711d563813a0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925870 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/293773da-e85d-43bc-b06a-19977c55c25a-trusted-ca\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925892 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-serving-cert\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925913 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8123393-4f9a-4546-8c76-10c87169eab3-config\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925953 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-serving-cert\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.925983 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-config\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.926019 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.927374 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c75ce80-6a91-407b-8035-97bc3149d046-auth-proxy-config\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.927918 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.927954 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.927963 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vsszz"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.926294 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bb59aca8-007d-46f0-a410-75fbe41c7217-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.928888 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-config\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930072 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bb59aca8-007d-46f0-a410-75fbe41c7217-audit-policies\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930648 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8gvh\" (UniqueName: \"kubernetes.io/projected/e48ec747-cb70-428b-9dfe-711d563813a0-kube-api-access-k8gvh\") pod \"openshift-apiserver-operator-796bbdcf4f-szg4v\" (UID: \"e48ec747-cb70-428b-9dfe-711d563813a0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930739 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8123393-4f9a-4546-8c76-10c87169eab3-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930771 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd2nq\" (UniqueName: \"kubernetes.io/projected/d8123393-4f9a-4546-8c76-10c87169eab3-kube-api-access-sd2nq\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930803 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-etcd-service-ca\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930826 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-oauth-config\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930860 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-trusted-ca-bundle\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930911 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bb59aca8-007d-46f0-a410-75fbe41c7217-encryption-config\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930938 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg8nc\" (UniqueName: \"kubernetes.io/projected/bb59aca8-007d-46f0-a410-75fbe41c7217-kube-api-access-gg8nc\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930976 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e33b0db2-705e-4424-985c-61d443a4cded-serving-cert\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930988 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e6d222cd-e154-421b-afb4-dbad33103c9a-images\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.930997 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931101 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-policies\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931103 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8123393-4f9a-4546-8c76-10c87169eab3-service-ca-bundle\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931133 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931154 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931172 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-client-ca\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931196 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgwdn\" (UniqueName: \"kubernetes.io/projected/e6d222cd-e154-421b-afb4-dbad33103c9a-kube-api-access-dgwdn\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931213 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c75ce80-6a91-407b-8035-97bc3149d046-config\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931216 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb59aca8-007d-46f0-a410-75fbe41c7217-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931259 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-dir\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931343 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-dir\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931589 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb59aca8-007d-46f0-a410-75fbe41c7217-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931636 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jtvv4"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.931657 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-console-config\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.932261 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-policies\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.932365 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6d222cd-e154-421b-afb4-dbad33103c9a-config\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.932498 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-gvrs8"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.932561 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.932824 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.932884 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.932970 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-gvrs8" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.933464 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.933707 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-config\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.933898 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-etcd-ca\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.933930 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-config\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.933977 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-service-ca\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.934257 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/293773da-e85d-43bc-b06a-19977c55c25a-config\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.935534 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/201c725b-a236-416d-8d93-2d0b21a103e0-metrics-tls\") pod \"dns-operator-744455d44c-hf8bj\" (UID: \"201c725b-a236-416d-8d93-2d0b21a103e0\") " pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.935983 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bb59aca8-007d-46f0-a410-75fbe41c7217-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.936168 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-client-ca\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.936170 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bb59aca8-007d-46f0-a410-75fbe41c7217-audit-dir\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.936299 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8123393-4f9a-4546-8c76-10c87169eab3-serving-cert\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.936973 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-trusted-ca-bundle\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.936994 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e48ec747-cb70-428b-9dfe-711d563813a0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-szg4v\" (UID: \"e48ec747-cb70-428b-9dfe-711d563813a0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.937081 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.937268 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e48ec747-cb70-428b-9dfe-711d563813a0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-szg4v\" (UID: \"e48ec747-cb70-428b-9dfe-711d563813a0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.937318 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8123393-4f9a-4546-8c76-10c87169eab3-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.937422 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8123393-4f9a-4546-8c76-10c87169eab3-config\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.937705 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/293773da-e85d-43bc-b06a-19977c55c25a-trusted-ca\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.938024 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-client-ca\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.938495 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.938806 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.939845 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.939974 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.940144 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-oauth-config\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.940223 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.940451 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.940812 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.940892 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e33b0db2-705e-4424-985c-61d443a4cded-serving-cert\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.941424 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb59aca8-007d-46f0-a410-75fbe41c7217-serving-cert\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.941766 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-oauth-serving-cert\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.941838 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.942235 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-serving-cert\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.942795 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fqjd6"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.943715 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.943740 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.944093 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/293773da-e85d-43bc-b06a-19977c55c25a-serving-cert\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.945040 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.945093 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.945402 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.945876 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.946114 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-serving-cert\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.946235 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.946270 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.946328 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6c75ce80-6a91-407b-8035-97bc3149d046-machine-approver-tls\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.947112 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-etcd-client\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.947318 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6d222cd-e154-421b-afb4-dbad33103c9a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.947637 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v5288"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.948635 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.948742 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bb59aca8-007d-46f0-a410-75fbe41c7217-encryption-config\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.949637 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.950553 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.951749 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jtvv4"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.952506 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-f9snq"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.953599 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.954521 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-gvrs8"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.955531 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.956510 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-68hsd"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.957325 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.957545 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-68hsd"] Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.960619 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb59aca8-007d-46f0-a410-75fbe41c7217-etcd-client\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.963032 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12954baa-8292-40c1-982a-28e23800078e-serving-cert\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.976342 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 08:54:38 crc kubenswrapper[4944]: I1124 08:54:38.995823 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.015923 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.035883 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.055754 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.076482 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.096234 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.115927 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.136144 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.155415 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.195846 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.215967 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.237619 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.256402 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.276483 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.296403 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.316280 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.342896 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.356288 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.376808 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.397397 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.417036 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.436478 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.457364 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.475433 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.497613 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.516772 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.537102 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.556622 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.576327 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.597454 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.616817 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.637093 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.657551 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.676354 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.706199 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.716719 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.738282 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.757567 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.777357 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.796118 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.816203 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.835070 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.854746 4944 request.go:700] Waited for 1.003854444s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.856727 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.875558 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.896175 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.916585 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.937342 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.956647 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.975717 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 08:54:39 crc kubenswrapper[4944]: I1124 08:54:39.995984 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.015973 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.037709 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.055521 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.076504 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.096088 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.117144 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.137363 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.156869 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.175686 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.196745 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.215809 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.237029 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.256574 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.276139 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.296014 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.315546 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.336014 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.356535 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.376202 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.396921 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.415995 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.435977 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.455285 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.476989 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.496635 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.516836 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.574393 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp7xs\" (UniqueName: \"kubernetes.io/projected/b355df00-f706-42c5-b16f-8c237a5048d8-kube-api-access-kp7xs\") pod \"console-f9d7485db-s9r2p\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.592280 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g79vv\" (UniqueName: \"kubernetes.io/projected/8ccd65fd-4f50-4351-8ac2-9a200ba85c6c-kube-api-access-g79vv\") pod \"etcd-operator-b45778765-dbf5z\" (UID: \"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.612430 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67n5z\" (UniqueName: \"kubernetes.io/projected/293773da-e85d-43bc-b06a-19977c55c25a-kube-api-access-67n5z\") pod \"console-operator-58897d9998-v2r8w\" (UID: \"293773da-e85d-43bc-b06a-19977c55c25a\") " pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.632108 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whtlb\" (UniqueName: \"kubernetes.io/projected/6c75ce80-6a91-407b-8035-97bc3149d046-kube-api-access-whtlb\") pod \"machine-approver-56656f9798-jqxxc\" (UID: \"6c75ce80-6a91-407b-8035-97bc3149d046\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.649132 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.655188 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms79n\" (UniqueName: \"kubernetes.io/projected/12954baa-8292-40c1-982a-28e23800078e-kube-api-access-ms79n\") pod \"route-controller-manager-6576b87f9c-v7pc7\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.672887 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sgvz\" (UniqueName: \"kubernetes.io/projected/e33b0db2-705e-4424-985c-61d443a4cded-kube-api-access-7sgvz\") pod \"controller-manager-879f6c89f-vsszz\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.676315 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.696025 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.715419 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.716710 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.733844 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.736140 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.756422 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.783377 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.790676 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.828155 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.830169 4944 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.838658 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg8nc\" (UniqueName: \"kubernetes.io/projected/bb59aca8-007d-46f0-a410-75fbe41c7217-kube-api-access-gg8nc\") pod \"apiserver-7bbb656c7d-ntscp\" (UID: \"bb59aca8-007d-46f0-a410-75fbe41c7217\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:40 crc kubenswrapper[4944]: W1124 08:54:40.841664 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c75ce80_6a91_407b_8035_97bc3149d046.slice/crio-ca103eae39b4be1d7209df995fe1e286cd737eb2bed5a710e844b408735ee7a4 WatchSource:0}: Error finding container ca103eae39b4be1d7209df995fe1e286cd737eb2bed5a710e844b408735ee7a4: Status 404 returned error can't find the container with id ca103eae39b4be1d7209df995fe1e286cd737eb2bed5a710e844b408735ee7a4 Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.860110 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgwdn\" (UniqueName: \"kubernetes.io/projected/e6d222cd-e154-421b-afb4-dbad33103c9a-kube-api-access-dgwdn\") pod \"machine-api-operator-5694c8668f-gg465\" (UID: \"e6d222cd-e154-421b-afb4-dbad33103c9a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.874135 4944 request.go:700] Waited for 1.937697875s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.876723 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ms8j\" (UniqueName: \"kubernetes.io/projected/aeedd47e-ece2-4717-839b-755f1e1cd8e0-kube-api-access-4ms8j\") pod \"oauth-openshift-558db77b4-h6vpv\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.895673 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-v2r8w"] Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.898731 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8gvh\" (UniqueName: \"kubernetes.io/projected/e48ec747-cb70-428b-9dfe-711d563813a0-kube-api-access-k8gvh\") pod \"openshift-apiserver-operator-796bbdcf4f-szg4v\" (UID: \"e48ec747-cb70-428b-9dfe-711d563813a0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.910181 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd2nq\" (UniqueName: \"kubernetes.io/projected/d8123393-4f9a-4546-8c76-10c87169eab3-kube-api-access-sd2nq\") pod \"authentication-operator-69f744f599-gs29j\" (UID: \"d8123393-4f9a-4546-8c76-10c87169eab3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.919786 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.934490 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spnpz\" (UniqueName: \"kubernetes.io/projected/201c725b-a236-416d-8d93-2d0b21a103e0-kube-api-access-spnpz\") pod \"dns-operator-744455d44c-hf8bj\" (UID: \"201c725b-a236-416d-8d93-2d0b21a103e0\") " pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.938103 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.955722 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-s9r2p"] Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.956530 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.976582 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 08:54:40 crc kubenswrapper[4944]: I1124 08:54:40.985973 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.002337 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dbf5z"] Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.006507 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.010412 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s9r2p" event={"ID":"b355df00-f706-42c5-b16f-8c237a5048d8","Type":"ContainerStarted","Data":"4782a3a25e0ad3ac5237b4047d6456430bc1b17cfee15ca912c0935dfd784881"} Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.012824 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" event={"ID":"6c75ce80-6a91-407b-8035-97bc3149d046","Type":"ContainerStarted","Data":"ca103eae39b4be1d7209df995fe1e286cd737eb2bed5a710e844b408735ee7a4"} Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.014480 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-v2r8w" event={"ID":"293773da-e85d-43bc-b06a-19977c55c25a","Type":"ContainerStarted","Data":"7ebb5b6ddfea24b536240696bf0f965be2728caceb8246eb064786c9657b3852"} Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.022657 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.061862 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8ssn\" (UniqueName: \"kubernetes.io/projected/3dc08e9e-07bc-4330-8caa-d73b9ac7faf4-kube-api-access-n8ssn\") pod \"openshift-config-operator-7777fb866f-x4nbt\" (UID: \"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.061904 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-tls\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.061924 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfbc9b3f-8917-4b9b-9892-1382156d7a56-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sn69z\" (UID: \"bfbc9b3f-8917-4b9b-9892-1382156d7a56\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.061947 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qls4\" (UniqueName: \"kubernetes.io/projected/bfbc9b3f-8917-4b9b-9892-1382156d7a56-kube-api-access-6qls4\") pod \"cluster-samples-operator-665b6dd947-sn69z\" (UID: \"bfbc9b3f-8917-4b9b-9892-1382156d7a56\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.061964 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/17c159e8-7de8-4716-b4cd-7f28a257a2bb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062000 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/69ba35de-8359-44fd-a9b0-e93e70f616eb-encryption-config\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062021 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-audit\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062037 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-etcd-serving-ca\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062052 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-image-import-ca\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062100 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062118 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcwnp\" (UniqueName: \"kubernetes.io/projected/69ba35de-8359-44fd-a9b0-e93e70f616eb-kube-api-access-kcwnp\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062139 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3dc08e9e-07bc-4330-8caa-d73b9ac7faf4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x4nbt\" (UID: \"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062163 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/69ba35de-8359-44fd-a9b0-e93e70f616eb-audit-dir\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062179 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/17c159e8-7de8-4716-b4cd-7f28a257a2bb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062227 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tms5r\" (UniqueName: \"kubernetes.io/projected/bbaf6605-39cb-41a0-900a-b2a87a00751e-kube-api-access-tms5r\") pod \"downloads-7954f5f757-9tglc\" (UID: \"bbaf6605-39cb-41a0-900a-b2a87a00751e\") " pod="openshift-console/downloads-7954f5f757-9tglc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062253 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/69ba35de-8359-44fd-a9b0-e93e70f616eb-etcd-client\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062269 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-bound-sa-token\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062285 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062302 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dc08e9e-07bc-4330-8caa-d73b9ac7faf4-serving-cert\") pod \"openshift-config-operator-7777fb866f-x4nbt\" (UID: \"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062318 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69ba35de-8359-44fd-a9b0-e93e70f616eb-serving-cert\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062334 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4wgl\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-kube-api-access-g4wgl\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062349 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/69ba35de-8359-44fd-a9b0-e93e70f616eb-node-pullsecrets\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062370 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-certificates\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062387 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-config\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.062408 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-trusted-ca\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.062534 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:41.562518707 +0000 UTC m=+142.096959169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.089315 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vsszz"] Nov 24 08:54:41 crc kubenswrapper[4944]: W1124 08:54:41.108368 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode33b0db2_705e_4424_985c_61d443a4cded.slice/crio-cc92af44ae7808071c687f1aebb8b4d0ec50a20afa833f7b18f9044b36139745 WatchSource:0}: Error finding container cc92af44ae7808071c687f1aebb8b4d0ec50a20afa833f7b18f9044b36139745: Status 404 returned error can't find the container with id cc92af44ae7808071c687f1aebb8b4d0ec50a20afa833f7b18f9044b36139745 Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.122874 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.163147 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.163577 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.163776 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-trusted-ca\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.163879 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfbc9b3f-8917-4b9b-9892-1382156d7a56-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sn69z\" (UID: \"bfbc9b3f-8917-4b9b-9892-1382156d7a56\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.163913 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:41.663873837 +0000 UTC m=+142.198314299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.163943 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/662092ff-6269-408c-8066-cbda91c4ecfe-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8m4x8\" (UID: \"662092ff-6269-408c-8066-cbda91c4ecfe\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.164011 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-tls\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.164156 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qls4\" (UniqueName: \"kubernetes.io/projected/bfbc9b3f-8917-4b9b-9892-1382156d7a56-kube-api-access-6qls4\") pod \"cluster-samples-operator-665b6dd947-sn69z\" (UID: \"bfbc9b3f-8917-4b9b-9892-1382156d7a56\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.164207 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/17c159e8-7de8-4716-b4cd-7f28a257a2bb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.164245 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/48f168d6-d783-45a5-851a-e0090dc07f7c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-d4q6h\" (UID: \"48f168d6-d783-45a5-851a-e0090dc07f7c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.164282 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjjj6\" (UniqueName: \"kubernetes.io/projected/745f9869-752f-4731-be2e-92ee56f613b0-kube-api-access-sjjj6\") pod \"openshift-controller-manager-operator-756b6f6bc6-4cm8w\" (UID: \"745f9869-752f-4731-be2e-92ee56f613b0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.166790 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/17c159e8-7de8-4716-b4cd-7f28a257a2bb-ca-trust-extracted\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.167575 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/22e1dda2-1c29-4f22-97c3-71c6aa588b47-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jz9hv\" (UID: \"22e1dda2-1c29-4f22-97c3-71c6aa588b47\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.167730 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtctv\" (UniqueName: \"kubernetes.io/projected/98143ac5-35b2-4689-872f-bf65e0e402e2-kube-api-access-mtctv\") pod \"machine-config-server-qwbqn\" (UID: \"98143ac5-35b2-4689-872f-bf65e0e402e2\") " pod="openshift-machine-config-operator/machine-config-server-qwbqn" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.167754 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/548a2d15-cb3c-43ec-9310-9661929b878e-default-certificate\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168571 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b98fb73-1a36-4740-8e2c-32802b4bc8f1-config\") pod \"service-ca-operator-777779d784-qbkqp\" (UID: \"1b98fb73-1a36-4740-8e2c-32802b4bc8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168678 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lg64\" (UniqueName: \"kubernetes.io/projected/8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0-kube-api-access-5lg64\") pod \"machine-config-controller-84d6567774-86wjt\" (UID: \"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168696 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-audit\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168732 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/798fde25-9b81-4d19-97e8-d9d953d58924-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168751 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168779 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/745f9869-752f-4731-be2e-92ee56f613b0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4cm8w\" (UID: \"745f9869-752f-4731-be2e-92ee56f613b0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168811 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0768e45f-6339-47f6-ba67-ffe401d2a502-secret-volume\") pod \"collect-profiles-29399565-9zv4s\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168830 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m6g58\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168858 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-etcd-serving-ca\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168912 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-image-import-ca\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168929 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-webhook-cert\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.168983 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6-cert\") pod \"ingress-canary-gvrs8\" (UID: \"1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6\") " pod="openshift-ingress-canary/ingress-canary-gvrs8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169011 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcwnp\" (UniqueName: \"kubernetes.io/projected/69ba35de-8359-44fd-a9b0-e93e70f616eb-kube-api-access-kcwnp\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169029 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d3d329-b103-4e77-9b5b-83535f5bbc44-config\") pod \"kube-apiserver-operator-766d6c64bb-bkwvc\" (UID: \"d4d3d329-b103-4e77-9b5b-83535f5bbc44\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169065 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d4d3d329-b103-4e77-9b5b-83535f5bbc44-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bkwvc\" (UID: \"d4d3d329-b103-4e77-9b5b-83535f5bbc44\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169129 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/98143ac5-35b2-4689-872f-bf65e0e402e2-certs\") pod \"machine-config-server-qwbqn\" (UID: \"98143ac5-35b2-4689-872f-bf65e0e402e2\") " pod="openshift-machine-config-operator/machine-config-server-qwbqn" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169171 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d3d329-b103-4e77-9b5b-83535f5bbc44-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bkwvc\" (UID: \"d4d3d329-b103-4e77-9b5b-83535f5bbc44\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169199 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzkhj\" (UniqueName: \"kubernetes.io/projected/22e1dda2-1c29-4f22-97c3-71c6aa588b47-kube-api-access-rzkhj\") pod \"control-plane-machine-set-operator-78cbb6b69f-jz9hv\" (UID: \"22e1dda2-1c29-4f22-97c3-71c6aa588b47\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169222 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-images\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169243 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/69ba35de-8359-44fd-a9b0-e93e70f616eb-audit-dir\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169279 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89r7n\" (UniqueName: \"kubernetes.io/projected/0f480216-c6e1-400b-90e2-91892fed172a-kube-api-access-89r7n\") pod \"service-ca-9c57cc56f-f9snq\" (UID: \"0f480216-c6e1-400b-90e2-91892fed172a\") " pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169317 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tms5r\" (UniqueName: \"kubernetes.io/projected/bbaf6605-39cb-41a0-900a-b2a87a00751e-kube-api-access-tms5r\") pod \"downloads-7954f5f757-9tglc\" (UID: \"bbaf6605-39cb-41a0-900a-b2a87a00751e\") " pod="openshift-console/downloads-7954f5f757-9tglc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169341 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5pnh\" (UniqueName: \"kubernetes.io/projected/449d38e0-1f3f-46fe-a256-1dbdc041900b-kube-api-access-j5pnh\") pod \"kube-storage-version-migrator-operator-b67b599dd-v48qh\" (UID: \"449d38e0-1f3f-46fe-a256-1dbdc041900b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169367 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-csi-data-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169392 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/81e802f8-50f6-4287-9960-4722c2299155-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fqjd6\" (UID: \"81e802f8-50f6-4287-9960-4722c2299155\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169418 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrpk7\" (UniqueName: \"kubernetes.io/projected/1b98fb73-1a36-4740-8e2c-32802b4bc8f1-kube-api-access-wrpk7\") pod \"service-ca-operator-777779d784-qbkqp\" (UID: \"1b98fb73-1a36-4740-8e2c-32802b4bc8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169441 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0-proxy-tls\") pod \"machine-config-controller-84d6567774-86wjt\" (UID: \"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169466 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65lwd\" (UniqueName: \"kubernetes.io/projected/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-kube-api-access-65lwd\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169488 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/548a2d15-cb3c-43ec-9310-9661929b878e-stats-auth\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169512 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx4gc\" (UniqueName: \"kubernetes.io/projected/48f168d6-d783-45a5-851a-e0090dc07f7c-kube-api-access-bx4gc\") pod \"olm-operator-6b444d44fb-d4q6h\" (UID: \"48f168d6-d783-45a5-851a-e0090dc07f7c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169535 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6tl5\" (UniqueName: \"kubernetes.io/projected/0768e45f-6339-47f6-ba67-ffe401d2a502-kube-api-access-c6tl5\") pod \"collect-profiles-29399565-9zv4s\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169568 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kq7r\" (UniqueName: \"kubernetes.io/projected/798fde25-9b81-4d19-97e8-d9d953d58924-kube-api-access-8kq7r\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169589 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlfkq\" (UniqueName: \"kubernetes.io/projected/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-kube-api-access-nlfkq\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169618 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-audit\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169625 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dc08e9e-07bc-4330-8caa-d73b9ac7faf4-serving-cert\") pod \"openshift-config-operator-7777fb866f-x4nbt\" (UID: \"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169683 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69ba35de-8359-44fd-a9b0-e93e70f616eb-serving-cert\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169714 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzffx\" (UniqueName: \"kubernetes.io/projected/5fcd0548-fb06-43f1-9328-a0f58b7db2da-kube-api-access-wzffx\") pod \"dns-default-68hsd\" (UID: \"5fcd0548-fb06-43f1-9328-a0f58b7db2da\") " pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169753 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4wgl\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-kube-api-access-g4wgl\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169774 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq9cj\" (UniqueName: \"kubernetes.io/projected/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-kube-api-access-pq9cj\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169806 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxz99\" (UniqueName: \"kubernetes.io/projected/807b8528-f6df-46ea-80f3-564a8e0a756a-kube-api-access-dxz99\") pod \"package-server-manager-789f6589d5-6j56z\" (UID: \"807b8528-f6df-46ea-80f3-564a8e0a756a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169827 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-certificates\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169843 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5fcd0548-fb06-43f1-9328-a0f58b7db2da-metrics-tls\") pod \"dns-default-68hsd\" (UID: \"5fcd0548-fb06-43f1-9328-a0f58b7db2da\") " pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169860 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f54d182a-01e5-4ac8-830a-3d73341d5099-profile-collector-cert\") pod \"catalog-operator-68c6474976-hwsml\" (UID: \"f54d182a-01e5-4ac8-830a-3d73341d5099\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169877 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trzhv\" (UniqueName: \"kubernetes.io/projected/81e802f8-50f6-4287-9960-4722c2299155-kube-api-access-trzhv\") pod \"multus-admission-controller-857f4d67dd-fqjd6\" (UID: \"81e802f8-50f6-4287-9960-4722c2299155\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169898 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-config\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169916 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-86wjt\" (UID: \"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169939 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/449d38e0-1f3f-46fe-a256-1dbdc041900b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-v48qh\" (UID: \"449d38e0-1f3f-46fe-a256-1dbdc041900b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169959 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-trusted-ca\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169980 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.169999 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8ssn\" (UniqueName: \"kubernetes.io/projected/3dc08e9e-07bc-4330-8caa-d73b9ac7faf4-kube-api-access-n8ssn\") pod \"openshift-config-operator-7777fb866f-x4nbt\" (UID: \"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170016 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-registration-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170040 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-socket-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170060 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/48f168d6-d783-45a5-851a-e0090dc07f7c-srv-cert\") pod \"olm-operator-6b444d44fb-d4q6h\" (UID: \"48f168d6-d783-45a5-851a-e0090dc07f7c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170170 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a322ced-c651-4f79-aeae-1fa31aea8fd0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6qvf\" (UID: \"2a322ced-c651-4f79-aeae-1fa31aea8fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170190 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-tmpfs\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170207 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/798fde25-9b81-4d19-97e8-d9d953d58924-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170252 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/69ba35de-8359-44fd-a9b0-e93e70f616eb-encryption-config\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170289 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0f480216-c6e1-400b-90e2-91892fed172a-signing-cabundle\") pod \"service-ca-9c57cc56f-f9snq\" (UID: \"0f480216-c6e1-400b-90e2-91892fed172a\") " pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170316 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/548a2d15-cb3c-43ec-9310-9661929b878e-metrics-certs\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170399 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170422 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/662092ff-6269-408c-8066-cbda91c4ecfe-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8m4x8\" (UID: \"662092ff-6269-408c-8066-cbda91c4ecfe\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170439 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f54d182a-01e5-4ac8-830a-3d73341d5099-srv-cert\") pod \"catalog-operator-68c6474976-hwsml\" (UID: \"f54d182a-01e5-4ac8-830a-3d73341d5099\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170458 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a322ced-c651-4f79-aeae-1fa31aea8fd0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6qvf\" (UID: \"2a322ced-c651-4f79-aeae-1fa31aea8fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170477 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/548a2d15-cb3c-43ec-9310-9661929b878e-service-ca-bundle\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170498 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0768e45f-6339-47f6-ba67-ffe401d2a502-config-volume\") pod \"collect-profiles-29399565-9zv4s\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170536 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fcd0548-fb06-43f1-9328-a0f58b7db2da-config-volume\") pod \"dns-default-68hsd\" (UID: \"5fcd0548-fb06-43f1-9328-a0f58b7db2da\") " pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170554 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj9bd\" (UniqueName: \"kubernetes.io/projected/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-kube-api-access-cj9bd\") pod \"marketplace-operator-79b997595-m6g58\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170569 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-mountpoint-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170588 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3dc08e9e-07bc-4330-8caa-d73b9ac7faf4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x4nbt\" (UID: \"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170607 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/17c159e8-7de8-4716-b4cd-7f28a257a2bb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170629 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0f480216-c6e1-400b-90e2-91892fed172a-signing-key\") pod \"service-ca-9c57cc56f-f9snq\" (UID: \"0f480216-c6e1-400b-90e2-91892fed172a\") " pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170658 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj5tr\" (UniqueName: \"kubernetes.io/projected/47dcb102-7868-4647-b85a-c7a6e44b66f5-kube-api-access-pj5tr\") pod \"migrator-59844c95c7-b7gc7\" (UID: \"47dcb102-7868-4647-b85a-c7a6e44b66f5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170677 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/449d38e0-1f3f-46fe-a256-1dbdc041900b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-v48qh\" (UID: \"449d38e0-1f3f-46fe-a256-1dbdc041900b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170692 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-apiservice-cert\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170722 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cmxn\" (UniqueName: \"kubernetes.io/projected/f54d182a-01e5-4ac8-830a-3d73341d5099-kube-api-access-4cmxn\") pod \"catalog-operator-68c6474976-hwsml\" (UID: \"f54d182a-01e5-4ac8-830a-3d73341d5099\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170738 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/98143ac5-35b2-4689-872f-bf65e0e402e2-node-bootstrap-token\") pod \"machine-config-server-qwbqn\" (UID: \"98143ac5-35b2-4689-872f-bf65e0e402e2\") " pod="openshift-machine-config-operator/machine-config-server-qwbqn" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170755 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/798fde25-9b81-4d19-97e8-d9d953d58924-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170775 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-proxy-tls\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170795 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-plugins-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170811 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xxht\" (UniqueName: \"kubernetes.io/projected/548a2d15-cb3c-43ec-9310-9661929b878e-kube-api-access-9xxht\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170828 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/69ba35de-8359-44fd-a9b0-e93e70f616eb-etcd-client\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170844 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m6g58\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170872 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-bound-sa-token\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170888 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170929 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b98fb73-1a36-4740-8e2c-32802b4bc8f1-serving-cert\") pod \"service-ca-operator-777779d784-qbkqp\" (UID: \"1b98fb73-1a36-4740-8e2c-32802b4bc8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170945 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a322ced-c651-4f79-aeae-1fa31aea8fd0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6qvf\" (UID: \"2a322ced-c651-4f79-aeae-1fa31aea8fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170962 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfdv8\" (UniqueName: \"kubernetes.io/projected/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-kube-api-access-sfdv8\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.170978 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/69ba35de-8359-44fd-a9b0-e93e70f616eb-node-pullsecrets\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.171082 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/662092ff-6269-408c-8066-cbda91c4ecfe-config\") pod \"kube-controller-manager-operator-78b949d7b-8m4x8\" (UID: \"662092ff-6269-408c-8066-cbda91c4ecfe\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.171101 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4nt5\" (UniqueName: \"kubernetes.io/projected/1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6-kube-api-access-g4nt5\") pod \"ingress-canary-gvrs8\" (UID: \"1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6\") " pod="openshift-ingress-canary/ingress-canary-gvrs8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.171116 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-metrics-tls\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.171132 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/745f9869-752f-4731-be2e-92ee56f613b0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4cm8w\" (UID: \"745f9869-752f-4731-be2e-92ee56f613b0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.171146 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/807b8528-f6df-46ea-80f3-564a8e0a756a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6j56z\" (UID: \"807b8528-f6df-46ea-80f3-564a8e0a756a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.173194 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/69ba35de-8359-44fd-a9b0-e93e70f616eb-audit-dir\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.173344 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-etcd-serving-ca\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.174062 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-config\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.175008 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/69ba35de-8359-44fd-a9b0-e93e70f616eb-node-pullsecrets\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.176889 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:41.676869763 +0000 UTC m=+142.211310425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.177464 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.177655 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-trusted-ca\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.178915 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-certificates\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.179354 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3dc08e9e-07bc-4330-8caa-d73b9ac7faf4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x4nbt\" (UID: \"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.182517 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7"] Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.183972 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/69ba35de-8359-44fd-a9b0-e93e70f616eb-image-import-ca\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.184623 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/bfbc9b3f-8917-4b9b-9892-1382156d7a56-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sn69z\" (UID: \"bfbc9b3f-8917-4b9b-9892-1382156d7a56\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.189183 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dc08e9e-07bc-4330-8caa-d73b9ac7faf4-serving-cert\") pod \"openshift-config-operator-7777fb866f-x4nbt\" (UID: \"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.189396 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/17c159e8-7de8-4716-b4cd-7f28a257a2bb-installation-pull-secrets\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.189754 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/69ba35de-8359-44fd-a9b0-e93e70f616eb-etcd-client\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.191898 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-tls\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.192418 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69ba35de-8359-44fd-a9b0-e93e70f616eb-serving-cert\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.199777 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/69ba35de-8359-44fd-a9b0-e93e70f616eb-encryption-config\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.207587 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.220585 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qls4\" (UniqueName: \"kubernetes.io/projected/bfbc9b3f-8917-4b9b-9892-1382156d7a56-kube-api-access-6qls4\") pod \"cluster-samples-operator-665b6dd947-sn69z\" (UID: \"bfbc9b3f-8917-4b9b-9892-1382156d7a56\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.236406 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.245360 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcwnp\" (UniqueName: \"kubernetes.io/projected/69ba35de-8359-44fd-a9b0-e93e70f616eb-kube-api-access-kcwnp\") pod \"apiserver-76f77b778f-tm2ss\" (UID: \"69ba35de-8359-44fd-a9b0-e93e70f616eb\") " pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.257810 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4wgl\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-kube-api-access-g4wgl\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.272149 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tms5r\" (UniqueName: \"kubernetes.io/projected/bbaf6605-39cb-41a0-900a-b2a87a00751e-kube-api-access-tms5r\") pod \"downloads-7954f5f757-9tglc\" (UID: \"bbaf6605-39cb-41a0-900a-b2a87a00751e\") " pod="openshift-console/downloads-7954f5f757-9tglc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.273379 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.273559 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f54d182a-01e5-4ac8-830a-3d73341d5099-srv-cert\") pod \"catalog-operator-68c6474976-hwsml\" (UID: \"f54d182a-01e5-4ac8-830a-3d73341d5099\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.273585 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/662092ff-6269-408c-8066-cbda91c4ecfe-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8m4x8\" (UID: \"662092ff-6269-408c-8066-cbda91c4ecfe\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.274135 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:41.774116301 +0000 UTC m=+142.308556763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274157 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a322ced-c651-4f79-aeae-1fa31aea8fd0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6qvf\" (UID: \"2a322ced-c651-4f79-aeae-1fa31aea8fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274178 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/548a2d15-cb3c-43ec-9310-9661929b878e-service-ca-bundle\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274197 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0768e45f-6339-47f6-ba67-ffe401d2a502-config-volume\") pod \"collect-profiles-29399565-9zv4s\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274528 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fcd0548-fb06-43f1-9328-a0f58b7db2da-config-volume\") pod \"dns-default-68hsd\" (UID: \"5fcd0548-fb06-43f1-9328-a0f58b7db2da\") " pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274546 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj9bd\" (UniqueName: \"kubernetes.io/projected/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-kube-api-access-cj9bd\") pod \"marketplace-operator-79b997595-m6g58\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274563 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-mountpoint-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274580 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0f480216-c6e1-400b-90e2-91892fed172a-signing-key\") pod \"service-ca-9c57cc56f-f9snq\" (UID: \"0f480216-c6e1-400b-90e2-91892fed172a\") " pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274599 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj5tr\" (UniqueName: \"kubernetes.io/projected/47dcb102-7868-4647-b85a-c7a6e44b66f5-kube-api-access-pj5tr\") pod \"migrator-59844c95c7-b7gc7\" (UID: \"47dcb102-7868-4647-b85a-c7a6e44b66f5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274617 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/449d38e0-1f3f-46fe-a256-1dbdc041900b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-v48qh\" (UID: \"449d38e0-1f3f-46fe-a256-1dbdc041900b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274632 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-apiservice-cert\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274648 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/798fde25-9b81-4d19-97e8-d9d953d58924-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274666 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cmxn\" (UniqueName: \"kubernetes.io/projected/f54d182a-01e5-4ac8-830a-3d73341d5099-kube-api-access-4cmxn\") pod \"catalog-operator-68c6474976-hwsml\" (UID: \"f54d182a-01e5-4ac8-830a-3d73341d5099\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274681 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/98143ac5-35b2-4689-872f-bf65e0e402e2-node-bootstrap-token\") pod \"machine-config-server-qwbqn\" (UID: \"98143ac5-35b2-4689-872f-bf65e0e402e2\") " pod="openshift-machine-config-operator/machine-config-server-qwbqn" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274700 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-plugins-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274716 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xxht\" (UniqueName: \"kubernetes.io/projected/548a2d15-cb3c-43ec-9310-9661929b878e-kube-api-access-9xxht\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274736 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-proxy-tls\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274757 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m6g58\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274779 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a322ced-c651-4f79-aeae-1fa31aea8fd0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6qvf\" (UID: \"2a322ced-c651-4f79-aeae-1fa31aea8fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274794 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfdv8\" (UniqueName: \"kubernetes.io/projected/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-kube-api-access-sfdv8\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274810 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b98fb73-1a36-4740-8e2c-32802b4bc8f1-serving-cert\") pod \"service-ca-operator-777779d784-qbkqp\" (UID: \"1b98fb73-1a36-4740-8e2c-32802b4bc8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274830 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/662092ff-6269-408c-8066-cbda91c4ecfe-config\") pod \"kube-controller-manager-operator-78b949d7b-8m4x8\" (UID: \"662092ff-6269-408c-8066-cbda91c4ecfe\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274847 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/745f9869-752f-4731-be2e-92ee56f613b0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4cm8w\" (UID: \"745f9869-752f-4731-be2e-92ee56f613b0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274864 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/807b8528-f6df-46ea-80f3-564a8e0a756a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6j56z\" (UID: \"807b8528-f6df-46ea-80f3-564a8e0a756a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274881 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4nt5\" (UniqueName: \"kubernetes.io/projected/1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6-kube-api-access-g4nt5\") pod \"ingress-canary-gvrs8\" (UID: \"1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6\") " pod="openshift-ingress-canary/ingress-canary-gvrs8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274896 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-metrics-tls\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274913 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-trusted-ca\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274945 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/662092ff-6269-408c-8066-cbda91c4ecfe-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8m4x8\" (UID: \"662092ff-6269-408c-8066-cbda91c4ecfe\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274981 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/48f168d6-d783-45a5-851a-e0090dc07f7c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-d4q6h\" (UID: \"48f168d6-d783-45a5-851a-e0090dc07f7c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.274998 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjjj6\" (UniqueName: \"kubernetes.io/projected/745f9869-752f-4731-be2e-92ee56f613b0-kube-api-access-sjjj6\") pod \"openshift-controller-manager-operator-756b6f6bc6-4cm8w\" (UID: \"745f9869-752f-4731-be2e-92ee56f613b0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275022 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/22e1dda2-1c29-4f22-97c3-71c6aa588b47-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jz9hv\" (UID: \"22e1dda2-1c29-4f22-97c3-71c6aa588b47\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275038 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b98fb73-1a36-4740-8e2c-32802b4bc8f1-config\") pod \"service-ca-operator-777779d784-qbkqp\" (UID: \"1b98fb73-1a36-4740-8e2c-32802b4bc8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275058 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtctv\" (UniqueName: \"kubernetes.io/projected/98143ac5-35b2-4689-872f-bf65e0e402e2-kube-api-access-mtctv\") pod \"machine-config-server-qwbqn\" (UID: \"98143ac5-35b2-4689-872f-bf65e0e402e2\") " pod="openshift-machine-config-operator/machine-config-server-qwbqn" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275089 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/548a2d15-cb3c-43ec-9310-9661929b878e-default-certificate\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275105 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lg64\" (UniqueName: \"kubernetes.io/projected/8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0-kube-api-access-5lg64\") pod \"machine-config-controller-84d6567774-86wjt\" (UID: \"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275123 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/798fde25-9b81-4d19-97e8-d9d953d58924-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275137 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275152 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/745f9869-752f-4731-be2e-92ee56f613b0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4cm8w\" (UID: \"745f9869-752f-4731-be2e-92ee56f613b0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275167 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0768e45f-6339-47f6-ba67-ffe401d2a502-secret-volume\") pod \"collect-profiles-29399565-9zv4s\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275182 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m6g58\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275204 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-webhook-cert\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275219 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6-cert\") pod \"ingress-canary-gvrs8\" (UID: \"1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6\") " pod="openshift-ingress-canary/ingress-canary-gvrs8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275235 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d3d329-b103-4e77-9b5b-83535f5bbc44-config\") pod \"kube-apiserver-operator-766d6c64bb-bkwvc\" (UID: \"d4d3d329-b103-4e77-9b5b-83535f5bbc44\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275250 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d4d3d329-b103-4e77-9b5b-83535f5bbc44-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bkwvc\" (UID: \"d4d3d329-b103-4e77-9b5b-83535f5bbc44\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275399 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/98143ac5-35b2-4689-872f-bf65e0e402e2-certs\") pod \"machine-config-server-qwbqn\" (UID: \"98143ac5-35b2-4689-872f-bf65e0e402e2\") " pod="openshift-machine-config-operator/machine-config-server-qwbqn" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275415 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d3d329-b103-4e77-9b5b-83535f5bbc44-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bkwvc\" (UID: \"d4d3d329-b103-4e77-9b5b-83535f5bbc44\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275432 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzkhj\" (UniqueName: \"kubernetes.io/projected/22e1dda2-1c29-4f22-97c3-71c6aa588b47-kube-api-access-rzkhj\") pod \"control-plane-machine-set-operator-78cbb6b69f-jz9hv\" (UID: \"22e1dda2-1c29-4f22-97c3-71c6aa588b47\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275446 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-images\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275464 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89r7n\" (UniqueName: \"kubernetes.io/projected/0f480216-c6e1-400b-90e2-91892fed172a-kube-api-access-89r7n\") pod \"service-ca-9c57cc56f-f9snq\" (UID: \"0f480216-c6e1-400b-90e2-91892fed172a\") " pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275483 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/81e802f8-50f6-4287-9960-4722c2299155-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fqjd6\" (UID: \"81e802f8-50f6-4287-9960-4722c2299155\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275499 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5pnh\" (UniqueName: \"kubernetes.io/projected/449d38e0-1f3f-46fe-a256-1dbdc041900b-kube-api-access-j5pnh\") pod \"kube-storage-version-migrator-operator-b67b599dd-v48qh\" (UID: \"449d38e0-1f3f-46fe-a256-1dbdc041900b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275514 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-csi-data-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275529 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrpk7\" (UniqueName: \"kubernetes.io/projected/1b98fb73-1a36-4740-8e2c-32802b4bc8f1-kube-api-access-wrpk7\") pod \"service-ca-operator-777779d784-qbkqp\" (UID: \"1b98fb73-1a36-4740-8e2c-32802b4bc8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275547 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0-proxy-tls\") pod \"machine-config-controller-84d6567774-86wjt\" (UID: \"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275564 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6tl5\" (UniqueName: \"kubernetes.io/projected/0768e45f-6339-47f6-ba67-ffe401d2a502-kube-api-access-c6tl5\") pod \"collect-profiles-29399565-9zv4s\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275579 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65lwd\" (UniqueName: \"kubernetes.io/projected/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-kube-api-access-65lwd\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275593 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/548a2d15-cb3c-43ec-9310-9661929b878e-stats-auth\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275609 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx4gc\" (UniqueName: \"kubernetes.io/projected/48f168d6-d783-45a5-851a-e0090dc07f7c-kube-api-access-bx4gc\") pod \"olm-operator-6b444d44fb-d4q6h\" (UID: \"48f168d6-d783-45a5-851a-e0090dc07f7c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275632 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kq7r\" (UniqueName: \"kubernetes.io/projected/798fde25-9b81-4d19-97e8-d9d953d58924-kube-api-access-8kq7r\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275653 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlfkq\" (UniqueName: \"kubernetes.io/projected/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-kube-api-access-nlfkq\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275676 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzffx\" (UniqueName: \"kubernetes.io/projected/5fcd0548-fb06-43f1-9328-a0f58b7db2da-kube-api-access-wzffx\") pod \"dns-default-68hsd\" (UID: \"5fcd0548-fb06-43f1-9328-a0f58b7db2da\") " pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275697 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq9cj\" (UniqueName: \"kubernetes.io/projected/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-kube-api-access-pq9cj\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275723 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxz99\" (UniqueName: \"kubernetes.io/projected/807b8528-f6df-46ea-80f3-564a8e0a756a-kube-api-access-dxz99\") pod \"package-server-manager-789f6589d5-6j56z\" (UID: \"807b8528-f6df-46ea-80f3-564a8e0a756a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275746 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5fcd0548-fb06-43f1-9328-a0f58b7db2da-metrics-tls\") pod \"dns-default-68hsd\" (UID: \"5fcd0548-fb06-43f1-9328-a0f58b7db2da\") " pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275768 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f54d182a-01e5-4ac8-830a-3d73341d5099-profile-collector-cert\") pod \"catalog-operator-68c6474976-hwsml\" (UID: \"f54d182a-01e5-4ac8-830a-3d73341d5099\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275790 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trzhv\" (UniqueName: \"kubernetes.io/projected/81e802f8-50f6-4287-9960-4722c2299155-kube-api-access-trzhv\") pod \"multus-admission-controller-857f4d67dd-fqjd6\" (UID: \"81e802f8-50f6-4287-9960-4722c2299155\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275810 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-86wjt\" (UID: \"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275831 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/449d38e0-1f3f-46fe-a256-1dbdc041900b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-v48qh\" (UID: \"449d38e0-1f3f-46fe-a256-1dbdc041900b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275850 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275868 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-registration-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275883 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-socket-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275898 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/48f168d6-d783-45a5-851a-e0090dc07f7c-srv-cert\") pod \"olm-operator-6b444d44fb-d4q6h\" (UID: \"48f168d6-d783-45a5-851a-e0090dc07f7c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275918 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a322ced-c651-4f79-aeae-1fa31aea8fd0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6qvf\" (UID: \"2a322ced-c651-4f79-aeae-1fa31aea8fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275942 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-tmpfs\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275963 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/798fde25-9b81-4d19-97e8-d9d953d58924-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.275988 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0f480216-c6e1-400b-90e2-91892fed172a-signing-cabundle\") pod \"service-ca-9c57cc56f-f9snq\" (UID: \"0f480216-c6e1-400b-90e2-91892fed172a\") " pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.276004 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/548a2d15-cb3c-43ec-9310-9661929b878e-metrics-certs\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.277200 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f54d182a-01e5-4ac8-830a-3d73341d5099-srv-cert\") pod \"catalog-operator-68c6474976-hwsml\" (UID: \"f54d182a-01e5-4ac8-830a-3d73341d5099\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.277747 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/662092ff-6269-408c-8066-cbda91c4ecfe-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8m4x8\" (UID: \"662092ff-6269-408c-8066-cbda91c4ecfe\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.278367 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a322ced-c651-4f79-aeae-1fa31aea8fd0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6qvf\" (UID: \"2a322ced-c651-4f79-aeae-1fa31aea8fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.279165 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-86wjt\" (UID: \"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.280610 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/548a2d15-cb3c-43ec-9310-9661929b878e-service-ca-bundle\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.281315 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0768e45f-6339-47f6-ba67-ffe401d2a502-config-volume\") pod \"collect-profiles-29399565-9zv4s\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.282547 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/449d38e0-1f3f-46fe-a256-1dbdc041900b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-v48qh\" (UID: \"449d38e0-1f3f-46fe-a256-1dbdc041900b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.282930 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fcd0548-fb06-43f1-9328-a0f58b7db2da-config-volume\") pod \"dns-default-68hsd\" (UID: \"5fcd0548-fb06-43f1-9328-a0f58b7db2da\") " pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.283147 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-mountpoint-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.283196 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a322ced-c651-4f79-aeae-1fa31aea8fd0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6qvf\" (UID: \"2a322ced-c651-4f79-aeae-1fa31aea8fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.283430 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/548a2d15-cb3c-43ec-9310-9661929b878e-metrics-certs\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.283839 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5fcd0548-fb06-43f1-9328-a0f58b7db2da-metrics-tls\") pod \"dns-default-68hsd\" (UID: \"5fcd0548-fb06-43f1-9328-a0f58b7db2da\") " pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.284465 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-registration-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.284481 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-socket-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.284469 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-tmpfs\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.285765 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0f480216-c6e1-400b-90e2-91892fed172a-signing-cabundle\") pod \"service-ca-9c57cc56f-f9snq\" (UID: \"0f480216-c6e1-400b-90e2-91892fed172a\") " pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.285886 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-csi-data-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.288790 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b98fb73-1a36-4740-8e2c-32802b4bc8f1-config\") pod \"service-ca-operator-777779d784-qbkqp\" (UID: \"1b98fb73-1a36-4740-8e2c-32802b4bc8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.288979 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-auth-proxy-config\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.289499 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/798fde25-9b81-4d19-97e8-d9d953d58924-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.289833 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/98143ac5-35b2-4689-872f-bf65e0e402e2-certs\") pod \"machine-config-server-qwbqn\" (UID: \"98143ac5-35b2-4689-872f-bf65e0e402e2\") " pod="openshift-machine-config-operator/machine-config-server-qwbqn" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.290157 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/449d38e0-1f3f-46fe-a256-1dbdc041900b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-v48qh\" (UID: \"449d38e0-1f3f-46fe-a256-1dbdc041900b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.290701 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0768e45f-6339-47f6-ba67-ffe401d2a502-secret-volume\") pod \"collect-profiles-29399565-9zv4s\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.291092 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/662092ff-6269-408c-8066-cbda91c4ecfe-config\") pod \"kube-controller-manager-operator-78b949d7b-8m4x8\" (UID: \"662092ff-6269-408c-8066-cbda91c4ecfe\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.291643 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/745f9869-752f-4731-be2e-92ee56f613b0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4cm8w\" (UID: \"745f9869-752f-4731-be2e-92ee56f613b0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.291934 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m6g58\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.292190 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/548a2d15-cb3c-43ec-9310-9661929b878e-default-certificate\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.292322 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-plugins-dir\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.294753 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d3d329-b103-4e77-9b5b-83535f5bbc44-config\") pod \"kube-apiserver-operator-766d6c64bb-bkwvc\" (UID: \"d4d3d329-b103-4e77-9b5b-83535f5bbc44\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.295221 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-trusted-ca\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.297268 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-webhook-cert\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.298099 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/807b8528-f6df-46ea-80f3-564a8e0a756a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6j56z\" (UID: \"807b8528-f6df-46ea-80f3-564a8e0a756a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.298343 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-metrics-tls\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.300059 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-images\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.300701 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/48f168d6-d783-45a5-851a-e0090dc07f7c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-d4q6h\" (UID: \"48f168d6-d783-45a5-851a-e0090dc07f7c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.304055 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/798fde25-9b81-4d19-97e8-d9d953d58924-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.304449 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/745f9869-752f-4731-be2e-92ee56f613b0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4cm8w\" (UID: \"745f9869-752f-4731-be2e-92ee56f613b0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.304929 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-proxy-tls\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.305123 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6-cert\") pod \"ingress-canary-gvrs8\" (UID: \"1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6\") " pod="openshift-ingress-canary/ingress-canary-gvrs8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.305198 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4d3d329-b103-4e77-9b5b-83535f5bbc44-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-bkwvc\" (UID: \"d4d3d329-b103-4e77-9b5b-83535f5bbc44\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.307527 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/22e1dda2-1c29-4f22-97c3-71c6aa588b47-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jz9hv\" (UID: \"22e1dda2-1c29-4f22-97c3-71c6aa588b47\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.308479 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/81e802f8-50f6-4287-9960-4722c2299155-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fqjd6\" (UID: \"81e802f8-50f6-4287-9960-4722c2299155\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.308733 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-apiservice-cert\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.309305 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0-proxy-tls\") pod \"machine-config-controller-84d6567774-86wjt\" (UID: \"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.309513 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b98fb73-1a36-4740-8e2c-32802b4bc8f1-serving-cert\") pod \"service-ca-operator-777779d784-qbkqp\" (UID: \"1b98fb73-1a36-4740-8e2c-32802b4bc8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.310158 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0f480216-c6e1-400b-90e2-91892fed172a-signing-key\") pod \"service-ca-9c57cc56f-f9snq\" (UID: \"0f480216-c6e1-400b-90e2-91892fed172a\") " pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.311730 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/48f168d6-d783-45a5-851a-e0090dc07f7c-srv-cert\") pod \"olm-operator-6b444d44fb-d4q6h\" (UID: \"48f168d6-d783-45a5-851a-e0090dc07f7c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.311903 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/548a2d15-cb3c-43ec-9310-9661929b878e-stats-auth\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.312322 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f54d182a-01e5-4ac8-830a-3d73341d5099-profile-collector-cert\") pod \"catalog-operator-68c6474976-hwsml\" (UID: \"f54d182a-01e5-4ac8-830a-3d73341d5099\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.312481 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/98143ac5-35b2-4689-872f-bf65e0e402e2-node-bootstrap-token\") pod \"machine-config-server-qwbqn\" (UID: \"98143ac5-35b2-4689-872f-bf65e0e402e2\") " pod="openshift-machine-config-operator/machine-config-server-qwbqn" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.314011 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-bound-sa-token\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.314966 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m6g58\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.326852 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8ssn\" (UniqueName: \"kubernetes.io/projected/3dc08e9e-07bc-4330-8caa-d73b9ac7faf4-kube-api-access-n8ssn\") pod \"openshift-config-operator-7777fb866f-x4nbt\" (UID: \"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.339676 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-gs29j"] Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.361663 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx4gc\" (UniqueName: \"kubernetes.io/projected/48f168d6-d783-45a5-851a-e0090dc07f7c-kube-api-access-bx4gc\") pod \"olm-operator-6b444d44fb-d4q6h\" (UID: \"48f168d6-d783-45a5-851a-e0090dc07f7c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.376964 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kq7r\" (UniqueName: \"kubernetes.io/projected/798fde25-9b81-4d19-97e8-d9d953d58924-kube-api-access-8kq7r\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.377357 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.377577 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp"] Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.378926 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:41.878906582 +0000 UTC m=+142.413347044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.399754 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gg465"] Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.402858 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlfkq\" (UniqueName: \"kubernetes.io/projected/f9ddea88-e412-4298-aa5c-fbd8f1938aa7-kube-api-access-nlfkq\") pod \"machine-config-operator-74547568cd-rvpws\" (UID: \"f9ddea88-e412-4298-aa5c-fbd8f1938aa7\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.407839 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:41 crc kubenswrapper[4944]: W1124 08:54:41.408375 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb59aca8_007d_46f0_a410_75fbe41c7217.slice/crio-6c721cf289f37858b8d8dd16ef2004a3c89a093f23da44dbbd520fce6b8ba8a0 WatchSource:0}: Error finding container 6c721cf289f37858b8d8dd16ef2004a3c89a093f23da44dbbd520fce6b8ba8a0: Status 404 returned error can't find the container with id 6c721cf289f37858b8d8dd16ef2004a3c89a093f23da44dbbd520fce6b8ba8a0 Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.413839 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzffx\" (UniqueName: \"kubernetes.io/projected/5fcd0548-fb06-43f1-9328-a0f58b7db2da-kube-api-access-wzffx\") pod \"dns-default-68hsd\" (UID: \"5fcd0548-fb06-43f1-9328-a0f58b7db2da\") " pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:41 crc kubenswrapper[4944]: W1124 08:54:41.434663 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6d222cd_e154_421b_afb4_dbad33103c9a.slice/crio-cf0922726bc3c359388c46d1d009ba8ad85fe6d89ba22c255169158a5563c8c1 WatchSource:0}: Error finding container cf0922726bc3c359388c46d1d009ba8ad85fe6d89ba22c255169158a5563c8c1: Status 404 returned error can't find the container with id cf0922726bc3c359388c46d1d009ba8ad85fe6d89ba22c255169158a5563c8c1 Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.437820 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq9cj\" (UniqueName: \"kubernetes.io/projected/9eca8c62-9cfd-44c9-ad1c-5898f00e8b48-kube-api-access-pq9cj\") pod \"packageserver-d55dfcdfc-86wfc\" (UID: \"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.451395 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.467623 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxz99\" (UniqueName: \"kubernetes.io/projected/807b8528-f6df-46ea-80f3-564a8e0a756a-kube-api-access-dxz99\") pod \"package-server-manager-789f6589d5-6j56z\" (UID: \"807b8528-f6df-46ea-80f3-564a8e0a756a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.476191 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.477913 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfdv8\" (UniqueName: \"kubernetes.io/projected/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-kube-api-access-sfdv8\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.478439 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.479136 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:41.979117459 +0000 UTC m=+142.513557921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.487487 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-h6vpv"] Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.499534 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89r7n\" (UniqueName: \"kubernetes.io/projected/0f480216-c6e1-400b-90e2-91892fed172a-kube-api-access-89r7n\") pod \"service-ca-9c57cc56f-f9snq\" (UID: \"0f480216-c6e1-400b-90e2-91892fed172a\") " pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.510503 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrpk7\" (UniqueName: \"kubernetes.io/projected/1b98fb73-1a36-4740-8e2c-32802b4bc8f1-kube-api-access-wrpk7\") pod \"service-ca-operator-777779d784-qbkqp\" (UID: \"1b98fb73-1a36-4740-8e2c-32802b4bc8f1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.524754 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.530141 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v"] Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.530887 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.536855 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj9bd\" (UniqueName: \"kubernetes.io/projected/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-kube-api-access-cj9bd\") pod \"marketplace-operator-79b997595-m6g58\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.541204 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.548633 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.558741 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjjj6\" (UniqueName: \"kubernetes.io/projected/745f9869-752f-4731-be2e-92ee56f613b0-kube-api-access-sjjj6\") pod \"openshift-controller-manager-operator-756b6f6bc6-4cm8w\" (UID: \"745f9869-752f-4731-be2e-92ee56f613b0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.564751 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.568410 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9tglc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.574581 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6tl5\" (UniqueName: \"kubernetes.io/projected/0768e45f-6339-47f6-ba67-ffe401d2a502-kube-api-access-c6tl5\") pod \"collect-profiles-29399565-9zv4s\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.582268 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.582681 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.082665094 +0000 UTC m=+142.617105556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.593847 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-hf8bj"] Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.594948 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.600610 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trzhv\" (UniqueName: \"kubernetes.io/projected/81e802f8-50f6-4287-9960-4722c2299155-kube-api-access-trzhv\") pod \"multus-admission-controller-857f4d67dd-fqjd6\" (UID: \"81e802f8-50f6-4287-9960-4722c2299155\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.623275 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0f91a43-15f8-425e-a5f3-9f5f15e3b506-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mjx2t\" (UID: \"d0f91a43-15f8-425e-a5f3-9f5f15e3b506\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: W1124 08:54:41.628482 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode48ec747_cb70_428b_9dfe_711d563813a0.slice/crio-283dd29bc212aa7940a9997ba742832a66262af6061bbf080f36fb33951d0da7 WatchSource:0}: Error finding container 283dd29bc212aa7940a9997ba742832a66262af6061bbf080f36fb33951d0da7: Status 404 returned error can't find the container with id 283dd29bc212aa7940a9997ba742832a66262af6061bbf080f36fb33951d0da7 Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.641341 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lg64\" (UniqueName: \"kubernetes.io/projected/8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0-kube-api-access-5lg64\") pod \"machine-config-controller-84d6567774-86wjt\" (UID: \"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" Nov 24 08:54:41 crc kubenswrapper[4944]: W1124 08:54:41.645492 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod201c725b_a236_416d_8d93_2d0b21a103e0.slice/crio-77b9ead155d6efb10070f5497d7c3d49c43f6f19e7d3d989c2ba3974af349292 WatchSource:0}: Error finding container 77b9ead155d6efb10070f5497d7c3d49c43f6f19e7d3d989c2ba3974af349292: Status 404 returned error can't find the container with id 77b9ead155d6efb10070f5497d7c3d49c43f6f19e7d3d989c2ba3974af349292 Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.658778 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/798fde25-9b81-4d19-97e8-d9d953d58924-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4t59q\" (UID: \"798fde25-9b81-4d19-97e8-d9d953d58924\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.681939 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5pnh\" (UniqueName: \"kubernetes.io/projected/449d38e0-1f3f-46fe-a256-1dbdc041900b-kube-api-access-j5pnh\") pod \"kube-storage-version-migrator-operator-b67b599dd-v48qh\" (UID: \"449d38e0-1f3f-46fe-a256-1dbdc041900b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.684657 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.684766 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.184743615 +0000 UTC m=+142.719184077 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.684937 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.685877 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z"] Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.686201 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.186181968 +0000 UTC m=+142.720622430 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.699400 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65lwd\" (UniqueName: \"kubernetes.io/projected/9448bfcf-6ec7-41ed-a6b5-f3243dec96af-kube-api-access-65lwd\") pod \"csi-hostpathplugin-jtvv4\" (UID: \"9448bfcf-6ec7-41ed-a6b5-f3243dec96af\") " pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.713669 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.720816 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d4d3d329-b103-4e77-9b5b-83535f5bbc44-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-bkwvc\" (UID: \"d4d3d329-b103-4e77-9b5b-83535f5bbc44\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.725897 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.734306 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.741997 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtctv\" (UniqueName: \"kubernetes.io/projected/98143ac5-35b2-4689-872f-bf65e0e402e2-kube-api-access-mtctv\") pod \"machine-config-server-qwbqn\" (UID: \"98143ac5-35b2-4689-872f-bf65e0e402e2\") " pod="openshift-machine-config-operator/machine-config-server-qwbqn" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.749253 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.754527 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.771206 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cmxn\" (UniqueName: \"kubernetes.io/projected/f54d182a-01e5-4ac8-830a-3d73341d5099-kube-api-access-4cmxn\") pod \"catalog-operator-68c6474976-hwsml\" (UID: \"f54d182a-01e5-4ac8-830a-3d73341d5099\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.776855 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4nt5\" (UniqueName: \"kubernetes.io/projected/1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6-kube-api-access-g4nt5\") pod \"ingress-canary-gvrs8\" (UID: \"1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6\") " pod="openshift-ingress-canary/ingress-canary-gvrs8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.784372 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.786370 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.790010 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.28998331 +0000 UTC m=+142.824423772 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.793441 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.802472 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.809124 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.816433 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qwbqn" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.819139 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj5tr\" (UniqueName: \"kubernetes.io/projected/47dcb102-7868-4647-b85a-c7a6e44b66f5-kube-api-access-pj5tr\") pod \"migrator-59844c95c7-b7gc7\" (UID: \"47dcb102-7868-4647-b85a-c7a6e44b66f5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.819427 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a322ced-c651-4f79-aeae-1fa31aea8fd0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w6qvf\" (UID: \"2a322ced-c651-4f79-aeae-1fa31aea8fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.837889 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tm2ss"] Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.845465 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xxht\" (UniqueName: \"kubernetes.io/projected/548a2d15-cb3c-43ec-9310-9661929b878e-kube-api-access-9xxht\") pod \"router-default-5444994796-bg5vl\" (UID: \"548a2d15-cb3c-43ec-9310-9661929b878e\") " pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.855891 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.868468 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/662092ff-6269-408c-8066-cbda91c4ecfe-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8m4x8\" (UID: \"662092ff-6269-408c-8066-cbda91c4ecfe\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.871139 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-gvrs8" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.879429 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzkhj\" (UniqueName: \"kubernetes.io/projected/22e1dda2-1c29-4f22-97c3-71c6aa588b47-kube-api-access-rzkhj\") pod \"control-plane-machine-set-operator-78cbb6b69f-jz9hv\" (UID: \"22e1dda2-1c29-4f22-97c3-71c6aa588b47\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.890811 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.891268 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.391252467 +0000 UTC m=+142.925692929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.891696 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.915354 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws"] Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.992588 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.992991 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.492968698 +0000 UTC m=+143.027409160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.993562 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:41 crc kubenswrapper[4944]: E1124 08:54:41.994444 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.494426932 +0000 UTC m=+143.028867394 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.997379 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:41 crc kubenswrapper[4944]: I1124 08:54:41.999616 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt"] Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.010186 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h"] Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.021645 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" Nov 24 08:54:42 crc kubenswrapper[4944]: W1124 08:54:42.026033 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69ba35de_8359_44fd_a9b0_e93e70f616eb.slice/crio-190a2ffe0765ff96be7b70859cd19d7ca933164b23e53ecd4f273fc0ccbf66f6 WatchSource:0}: Error finding container 190a2ffe0765ff96be7b70859cd19d7ca933164b23e53ecd4f273fc0ccbf66f6: Status 404 returned error can't find the container with id 190a2ffe0765ff96be7b70859cd19d7ca933164b23e53ecd4f273fc0ccbf66f6 Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.032700 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-v2r8w" event={"ID":"293773da-e85d-43bc-b06a-19977c55c25a","Type":"ContainerStarted","Data":"75b9a60769561cc6d40e0d603cccd07bb9629a9a0d21c139245c852692a0e6ca"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.033765 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.036800 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" event={"ID":"12954baa-8292-40c1-982a-28e23800078e","Type":"ContainerStarted","Data":"6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.036826 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" event={"ID":"12954baa-8292-40c1-982a-28e23800078e","Type":"ContainerStarted","Data":"3007e310db599d5ce9c596a3060d492ffaf7ca900b29c9727de39d4d79f52453"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.038401 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.038608 4944 patch_prober.go:28] interesting pod/console-operator-58897d9998-v2r8w container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.038646 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-v2r8w" podUID="293773da-e85d-43bc-b06a-19977c55c25a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.042876 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.043484 4944 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-v7pc7 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.043561 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" podUID="12954baa-8292-40c1-982a-28e23800078e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.045726 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" event={"ID":"e33b0db2-705e-4424-985c-61d443a4cded","Type":"ContainerStarted","Data":"de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.045783 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" event={"ID":"e33b0db2-705e-4424-985c-61d443a4cded","Type":"ContainerStarted","Data":"cc92af44ae7808071c687f1aebb8b4d0ec50a20afa833f7b18f9044b36139745"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.046419 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.048327 4944 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-vsszz container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.048404 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" podUID="e33b0db2-705e-4424-985c-61d443a4cded" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.054783 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" event={"ID":"e48ec747-cb70-428b-9dfe-711d563813a0","Type":"ContainerStarted","Data":"283dd29bc212aa7940a9997ba742832a66262af6061bbf080f36fb33951d0da7"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.062479 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7" Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.069411 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv" Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.074475 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" event={"ID":"d8123393-4f9a-4546-8c76-10c87169eab3","Type":"ContainerStarted","Data":"3bd3030e4ed0ea8472131f4621132250724b33542f362eab786969b754f12e71"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.074530 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" event={"ID":"d8123393-4f9a-4546-8c76-10c87169eab3","Type":"ContainerStarted","Data":"c275a02cf2ef0e0606223d496fb5c760ea5a7531e425fd37c405114738d93c2d"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.085744 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" event={"ID":"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c","Type":"ContainerStarted","Data":"66623c3e4d9888699fccdf3b755d4e1b9226f3103216c578e3699b6e36607cef"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.085796 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" event={"ID":"8ccd65fd-4f50-4351-8ac2-9a200ba85c6c","Type":"ContainerStarted","Data":"d18a3981c0a908990c563e53a519bbaf0af2ddcd5d98f290eb61095b0f95dea9"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.088773 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s9r2p" event={"ID":"b355df00-f706-42c5-b16f-8c237a5048d8","Type":"ContainerStarted","Data":"44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.092523 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" event={"ID":"bfbc9b3f-8917-4b9b-9892-1382156d7a56","Type":"ContainerStarted","Data":"56e1d3b6d27ff246d3c282a91bc6aaa8416f64bfe90ba2b445cc5cb0dffea93c"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.094660 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.094907 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.594881955 +0000 UTC m=+143.129322417 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.094975 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.096032 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.596010698 +0000 UTC m=+143.130451160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.101591 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-f9snq"] Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.109299 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" event={"ID":"bb59aca8-007d-46f0-a410-75fbe41c7217","Type":"ContainerStarted","Data":"6c721cf289f37858b8d8dd16ef2004a3c89a093f23da44dbbd520fce6b8ba8a0"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.111719 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" event={"ID":"aeedd47e-ece2-4717-839b-755f1e1cd8e0","Type":"ContainerStarted","Data":"aad46a88bb352e8961b434bef4f5404cae3a0d06e77631ac5c7ab3c1eff7a003"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.119089 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" event={"ID":"201c725b-a236-416d-8d93-2d0b21a103e0","Type":"ContainerStarted","Data":"77b9ead155d6efb10070f5497d7c3d49c43f6f19e7d3d989c2ba3974af349292"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.138798 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" event={"ID":"e6d222cd-e154-421b-afb4-dbad33103c9a","Type":"ContainerStarted","Data":"ff438bd0fa3d5ee1d52c2203917590ffd09bd5877fa78c4d5147c67e22676172"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.139389 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" event={"ID":"e6d222cd-e154-421b-afb4-dbad33103c9a","Type":"ContainerStarted","Data":"cf0922726bc3c359388c46d1d009ba8ad85fe6d89ba22c255169158a5563c8c1"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.176028 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" event={"ID":"6c75ce80-6a91-407b-8035-97bc3149d046","Type":"ContainerStarted","Data":"b3907a6d223552fad51e0d766890e9d9e9bfcab26d3a93ea85b764dec6b98764"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.176122 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" event={"ID":"6c75ce80-6a91-407b-8035-97bc3149d046","Type":"ContainerStarted","Data":"c941050333d6576185a10b17cb73c47580e97a5ce99d48e766f149c6588e1c23"} Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.196729 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.196937 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.696905024 +0000 UTC m=+143.231345486 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.197317 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.205217 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.705187429 +0000 UTC m=+143.239627891 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.230463 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp"] Nov 24 08:54:42 crc kubenswrapper[4944]: W1124 08:54:42.254011 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f480216_c6e1_400b_90e2_91892fed172a.slice/crio-87db6d3c566af6a1f82a8a7f598436c5fd53016139cfc7c7be0011753f2da8ed WatchSource:0}: Error finding container 87db6d3c566af6a1f82a8a7f598436c5fd53016139cfc7c7be0011753f2da8ed: Status 404 returned error can't find the container with id 87db6d3c566af6a1f82a8a7f598436c5fd53016139cfc7c7be0011753f2da8ed Nov 24 08:54:42 crc kubenswrapper[4944]: W1124 08:54:42.278290 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48f168d6_d783_45a5_851a_e0090dc07f7c.slice/crio-5c16363ec2714166b205a78e73397a7fd2aa5cdc830522e20382a0471225c409 WatchSource:0}: Error finding container 5c16363ec2714166b205a78e73397a7fd2aa5cdc830522e20382a0471225c409: Status 404 returned error can't find the container with id 5c16363ec2714166b205a78e73397a7fd2aa5cdc830522e20382a0471225c409 Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.302637 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.303862 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.803843349 +0000 UTC m=+143.338283811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.406446 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.407025 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:42.907009943 +0000 UTC m=+143.441450395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.507374 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.508139 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:43.008120946 +0000 UTC m=+143.542561408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.610498 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.610966 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:43.110950409 +0000 UTC m=+143.645390871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.639579 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9tglc"] Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.660135 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc"] Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.711205 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.712378 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:43.212356031 +0000 UTC m=+143.746796503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.817610 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.818053 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:43.318001288 +0000 UTC m=+143.852441760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.911270 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-68hsd"] Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.920641 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.921022 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:43.420990977 +0000 UTC m=+143.955431439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:42 crc kubenswrapper[4944]: I1124 08:54:42.921991 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:42 crc kubenswrapper[4944]: E1124 08:54:42.922605 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:43.422590354 +0000 UTC m=+143.957030826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.025646 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:43 crc kubenswrapper[4944]: E1124 08:54:43.026523 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:43.526501199 +0000 UTC m=+144.060941662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.128511 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:43 crc kubenswrapper[4944]: E1124 08:54:43.129003 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:43.628990533 +0000 UTC m=+144.163430985 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.187565 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" event={"ID":"aeedd47e-ece2-4717-839b-755f1e1cd8e0","Type":"ContainerStarted","Data":"1d7bcf30759d7b3a498a3f37641862f62b03336605dfd8af6b8a78981a692c83"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.191853 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.193285 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qwbqn" event={"ID":"98143ac5-35b2-4689-872f-bf65e0e402e2","Type":"ContainerStarted","Data":"2f6547d35cc450bdeae894b27db6932465b13983d18cdbe2fcfc48fbd996ae8e"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.196129 4944 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-h6vpv container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.6:6443/healthz\": dial tcp 10.217.0.6:6443: connect: connection refused" start-of-body= Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.196186 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" podUID="aeedd47e-ece2-4717-839b-755f1e1cd8e0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.6:6443/healthz\": dial tcp 10.217.0.6:6443: connect: connection refused" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.196573 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" event={"ID":"e48ec747-cb70-428b-9dfe-711d563813a0","Type":"ContainerStarted","Data":"76014fe1104cb487f6c2b147534bfad0b5821f7e0ed65f2cd115852e69218b4e"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.200847 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" event={"ID":"bfbc9b3f-8917-4b9b-9892-1382156d7a56","Type":"ContainerStarted","Data":"e2248cd814947c96e5344e8664994d12055b73fcce20cb172b02f55e3ad6d7b1"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.202812 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9tglc" event={"ID":"bbaf6605-39cb-41a0-900a-b2a87a00751e","Type":"ContainerStarted","Data":"9c0a1bb07e7f816e40ff10f712e6ff47b16395dc30db34de20837b3beef0e83d"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.205524 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" event={"ID":"48f168d6-d783-45a5-851a-e0090dc07f7c","Type":"ContainerStarted","Data":"5c16363ec2714166b205a78e73397a7fd2aa5cdc830522e20382a0471225c409"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.211097 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" event={"ID":"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48","Type":"ContainerStarted","Data":"bded19e3d9a67f0086036f8915ae41cba751eafb017233b8fc85b951031eac28"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.221132 4944 generic.go:334] "Generic (PLEG): container finished" podID="bb59aca8-007d-46f0-a410-75fbe41c7217" containerID="ae0b5feca659e94ce5363a8a2193b52fa2c4c2c6cebdd6d3090251bf242d0962" exitCode=0 Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.222354 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" event={"ID":"bb59aca8-007d-46f0-a410-75fbe41c7217","Type":"ContainerDied","Data":"ae0b5feca659e94ce5363a8a2193b52fa2c4c2c6cebdd6d3090251bf242d0962"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.228855 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" podStartSLOduration=124.228830679 podStartE2EDuration="2m4.228830679s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:43.225988784 +0000 UTC m=+143.760429246" watchObservedRunningTime="2025-11-24 08:54:43.228830679 +0000 UTC m=+143.763271141" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.229643 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:43 crc kubenswrapper[4944]: E1124 08:54:43.230039 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:43.730023183 +0000 UTC m=+144.264463645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.234658 4944 generic.go:334] "Generic (PLEG): container finished" podID="3dc08e9e-07bc-4330-8caa-d73b9ac7faf4" containerID="4a0cf9e441604a4f7a49564587fe35058a016c0b99e61117a2bdb42e6c7c53ea" exitCode=0 Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.234796 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" event={"ID":"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4","Type":"ContainerDied","Data":"4a0cf9e441604a4f7a49564587fe35058a016c0b99e61117a2bdb42e6c7c53ea"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.234843 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" event={"ID":"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4","Type":"ContainerStarted","Data":"79a77bc13bd7ae43fb1b1f4eaea2ec3f7c1f502cc7f32ab297a425e7b1835318"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.236869 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" event={"ID":"1b98fb73-1a36-4740-8e2c-32802b4bc8f1","Type":"ContainerStarted","Data":"9a9fb81fb61b56d3c29ab62a845ff1788b2310921557225f961873248d681250"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.237867 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bg5vl" event={"ID":"548a2d15-cb3c-43ec-9310-9661929b878e","Type":"ContainerStarted","Data":"ad6930bde316c0b8ba63b17adb556c73c5a7475fd2dded3b3c4ebfa27ef3d6de"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.243100 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" event={"ID":"0f480216-c6e1-400b-90e2-91892fed172a","Type":"ContainerStarted","Data":"87db6d3c566af6a1f82a8a7f598436c5fd53016139cfc7c7be0011753f2da8ed"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.255071 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" event={"ID":"69ba35de-8359-44fd-a9b0-e93e70f616eb","Type":"ContainerStarted","Data":"190a2ffe0765ff96be7b70859cd19d7ca933164b23e53ecd4f273fc0ccbf66f6"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.286852 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" event={"ID":"e6d222cd-e154-421b-afb4-dbad33103c9a","Type":"ContainerStarted","Data":"cdba47dda0f301388ef38f28e6cb299c5bebccbb8cf63e818a925e3e82717b03"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.296501 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" event={"ID":"f9ddea88-e412-4298-aa5c-fbd8f1938aa7","Type":"ContainerStarted","Data":"c170d5aa1ed158e84f5c45b30e7b912c19c75ac8c1296f1c26915239d2419f31"} Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.309904 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-gs29j" podStartSLOduration=124.309870537 podStartE2EDuration="2m4.309870537s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:43.296952534 +0000 UTC m=+143.831393026" watchObservedRunningTime="2025-11-24 08:54:43.309870537 +0000 UTC m=+143.844310999" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.319690 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.320723 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.323313 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-v2r8w" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.334792 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:43 crc kubenswrapper[4944]: E1124 08:54:43.337103 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:43.837040401 +0000 UTC m=+144.371480863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.344481 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" podStartSLOduration=124.34443855 podStartE2EDuration="2m4.34443855s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:43.338157394 +0000 UTC m=+143.872597866" watchObservedRunningTime="2025-11-24 08:54:43.34443855 +0000 UTC m=+143.878879022" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.436507 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.438435 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-v2r8w" podStartSLOduration=124.438419562 podStartE2EDuration="2m4.438419562s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:43.428403545 +0000 UTC m=+143.962844007" watchObservedRunningTime="2025-11-24 08:54:43.438419562 +0000 UTC m=+143.972860024" Nov 24 08:54:43 crc kubenswrapper[4944]: E1124 08:54:43.439823 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:43.939790852 +0000 UTC m=+144.474231514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.519342 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jqxxc" podStartSLOduration=124.519300116 podStartE2EDuration="2m4.519300116s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:43.45862842 +0000 UTC m=+143.993068902" watchObservedRunningTime="2025-11-24 08:54:43.519300116 +0000 UTC m=+144.053740578" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.557663 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:43 crc kubenswrapper[4944]: E1124 08:54:43.563890 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.063855415 +0000 UTC m=+144.598295887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.576628 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z"] Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.628600 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-s9r2p" podStartSLOduration=124.62856555 podStartE2EDuration="2m4.62856555s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:43.569140021 +0000 UTC m=+144.103580493" watchObservedRunningTime="2025-11-24 08:54:43.62856555 +0000 UTC m=+144.163006012" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.633009 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc"] Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.645320 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fqjd6"] Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.649329 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q"] Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.658905 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:43 crc kubenswrapper[4944]: E1124 08:54:43.659475 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.159455434 +0000 UTC m=+144.693895896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.660714 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m6g58"] Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.679251 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh"] Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.684398 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-dbf5z" podStartSLOduration=124.684375572 podStartE2EDuration="2m4.684375572s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:43.658427214 +0000 UTC m=+144.192867676" watchObservedRunningTime="2025-11-24 08:54:43.684375572 +0000 UTC m=+144.218816024" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.760905 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:43 crc kubenswrapper[4944]: E1124 08:54:43.762638 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.262618698 +0000 UTC m=+144.797059170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.845206 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-szg4v" podStartSLOduration=124.845174841 podStartE2EDuration="2m4.845174841s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:43.802510179 +0000 UTC m=+144.336950651" watchObservedRunningTime="2025-11-24 08:54:43.845174841 +0000 UTC m=+144.379615303" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.866033 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:43 crc kubenswrapper[4944]: E1124 08:54:43.866189 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.366157232 +0000 UTC m=+144.900597704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.866418 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:43 crc kubenswrapper[4944]: E1124 08:54:43.866832 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.366817422 +0000 UTC m=+144.901257884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.870353 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf"] Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.897840 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" podStartSLOduration=124.89781555 podStartE2EDuration="2m4.89781555s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:43.892209544 +0000 UTC m=+144.426650006" watchObservedRunningTime="2025-11-24 08:54:43.89781555 +0000 UTC m=+144.432256012" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.913232 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt"] Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.967551 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:43 crc kubenswrapper[4944]: E1124 08:54:43.968086 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.468035968 +0000 UTC m=+145.002476430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.976818 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-gg465" podStartSLOduration=124.976792527 podStartE2EDuration="2m4.976792527s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:43.944435999 +0000 UTC m=+144.478876471" watchObservedRunningTime="2025-11-24 08:54:43.976792527 +0000 UTC m=+144.511233009" Nov 24 08:54:43 crc kubenswrapper[4944]: I1124 08:54:43.977840 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w"] Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:43.988522 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t"] Nov 24 08:54:44 crc kubenswrapper[4944]: W1124 08:54:43.998424 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8868f76b_cfe5_4fb4_8866_e0b5ee27d3f0.slice/crio-fa6d16920c431c04c1484b3f5c62174cee968fd513ae62811976c67c2347b6fc WatchSource:0}: Error finding container fa6d16920c431c04c1484b3f5c62174cee968fd513ae62811976c67c2347b6fc: Status 404 returned error can't find the container with id fa6d16920c431c04c1484b3f5c62174cee968fd513ae62811976c67c2347b6fc Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.027192 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7"] Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.072819 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml"] Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.076237 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:44 crc kubenswrapper[4944]: E1124 08:54:44.076787 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.576766956 +0000 UTC m=+145.111207418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.094307 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s"] Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.113355 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-gvrs8"] Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.140831 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jtvv4"] Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.158510 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv"] Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.178642 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:44 crc kubenswrapper[4944]: E1124 08:54:44.179420 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.679392754 +0000 UTC m=+145.213833216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.179600 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:44 crc kubenswrapper[4944]: E1124 08:54:44.193173 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.69312415 +0000 UTC m=+145.227564612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.220479 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8"] Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.281627 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:44 crc kubenswrapper[4944]: E1124 08:54:44.282188 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.782112574 +0000 UTC m=+145.316553026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.337365 4944 generic.go:334] "Generic (PLEG): container finished" podID="69ba35de-8359-44fd-a9b0-e93e70f616eb" containerID="cbac2d7501cebfa2c49419e2a56f390d1d15e6ae05acc4794d00641eee44b038" exitCode=0 Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.337453 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" event={"ID":"69ba35de-8359-44fd-a9b0-e93e70f616eb","Type":"ContainerDied","Data":"cbac2d7501cebfa2c49419e2a56f390d1d15e6ae05acc4794d00641eee44b038"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.371484 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9tglc" event={"ID":"bbaf6605-39cb-41a0-900a-b2a87a00751e","Type":"ContainerStarted","Data":"1f77d82865ecce4eddc1bc92fba0cafd2de17632b48efbd96ac9a25e31e89eb6"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.372927 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-9tglc" Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.385037 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:44 crc kubenswrapper[4944]: E1124 08:54:44.385465 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.885450652 +0000 UTC m=+145.419891114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.408704 4944 patch_prober.go:28] interesting pod/downloads-7954f5f757-9tglc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.408787 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9tglc" podUID="bbaf6605-39cb-41a0-900a-b2a87a00751e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.486101 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:44 crc kubenswrapper[4944]: E1124 08:54:44.487834 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:44.987812283 +0000 UTC m=+145.522252755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.517509 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" event={"ID":"d4d3d329-b103-4e77-9b5b-83535f5bbc44","Type":"ContainerStarted","Data":"686ba38807b2b1e78a9e57c414c9dd864fe4a1b19ce2f0c8f7a0c28fdd609d10"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.524574 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" event={"ID":"81e802f8-50f6-4287-9960-4722c2299155","Type":"ContainerStarted","Data":"934f42207af4df369725a869f120e4ed8d1bbc09a81a0ae4b14d3982d3c3a721"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.575498 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" event={"ID":"807b8528-f6df-46ea-80f3-564a8e0a756a","Type":"ContainerStarted","Data":"d5499e256c9e8fcc3732ab28e077657df2a4fb482119ab5dd8bdc9bee01b6e9e"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.589526 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:44 crc kubenswrapper[4944]: E1124 08:54:44.590189 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:45.090176352 +0000 UTC m=+145.624616814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.629533 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" event={"ID":"0f480216-c6e1-400b-90e2-91892fed172a","Type":"ContainerStarted","Data":"47b8d355dcbc0df216d05cfba0e8dd3f125f24c35ef477bcdcfc0f68cc1370eb"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.638413 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" event={"ID":"46615e45-27d1-47b4-a5d7-f0cfb9f7476d","Type":"ContainerStarted","Data":"76c9207008b198cb2271af30032a9a9e2dd8d143bf18ec75883776947daccddf"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.654265 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-f9snq" podStartSLOduration=125.654238348 podStartE2EDuration="2m5.654238348s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:44.652946581 +0000 UTC m=+145.187387073" watchObservedRunningTime="2025-11-24 08:54:44.654238348 +0000 UTC m=+145.188678810" Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.655247 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-9tglc" podStartSLOduration=125.655238618 podStartE2EDuration="2m5.655238618s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:44.467972935 +0000 UTC m=+145.002413407" watchObservedRunningTime="2025-11-24 08:54:44.655238618 +0000 UTC m=+145.189679080" Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.692613 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:44 crc kubenswrapper[4944]: E1124 08:54:44.693605 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:45.193567522 +0000 UTC m=+145.728007984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.744599 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" podStartSLOduration=125.744576262 podStartE2EDuration="2m5.744576262s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:44.742672126 +0000 UTC m=+145.277112588" watchObservedRunningTime="2025-11-24 08:54:44.744576262 +0000 UTC m=+145.279016724" Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.750433 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" event={"ID":"3dc08e9e-07bc-4330-8caa-d73b9ac7faf4","Type":"ContainerStarted","Data":"90f1db16e8962eb1edd7005d9c8907f4f626d17798e9e804293627cb5505a6fc"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.750583 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.788598 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" event={"ID":"745f9869-752f-4731-be2e-92ee56f613b0","Type":"ContainerStarted","Data":"88c06a7c4d9e624016650369cdd760918e21aa1517568303b80375b31a3290bc"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.794065 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:44 crc kubenswrapper[4944]: E1124 08:54:44.794426 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:45.294412797 +0000 UTC m=+145.828853259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.838225 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-68hsd" event={"ID":"5fcd0548-fb06-43f1-9328-a0f58b7db2da","Type":"ContainerStarted","Data":"cedddd591a84307f25d45065193cfb1087cf0eba5f90a207a0ce2175bc7ffe30"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.873766 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" event={"ID":"1b98fb73-1a36-4740-8e2c-32802b4bc8f1","Type":"ContainerStarted","Data":"005218f1516b932e749460a3a41f8ff5138b6f62a4f8476839b9cb236c52ee87"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.895824 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:44 crc kubenswrapper[4944]: E1124 08:54:44.896220 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:45.39618752 +0000 UTC m=+145.930628142 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.897693 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" event={"ID":"bfbc9b3f-8917-4b9b-9892-1382156d7a56","Type":"ContainerStarted","Data":"b333780841d7fc443dc46548e209c8d503233988d95eb8692b121a11ddcd808b"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.907735 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" podStartSLOduration=125.907711271 podStartE2EDuration="2m5.907711271s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:44.809944368 +0000 UTC m=+145.344384830" watchObservedRunningTime="2025-11-24 08:54:44.907711271 +0000 UTC m=+145.442151733" Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.921153 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" event={"ID":"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0","Type":"ContainerStarted","Data":"fa6d16920c431c04c1484b3f5c62174cee968fd513ae62811976c67c2347b6fc"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.935189 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bg5vl" event={"ID":"548a2d15-cb3c-43ec-9310-9661929b878e","Type":"ContainerStarted","Data":"3b6f0bb21a79f0d6f1827ffe37e4dfacc5c0bba2ea5fed484eb0b15af0f37fed"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.958586 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qbkqp" podStartSLOduration=125.958531316 podStartE2EDuration="2m5.958531316s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:44.913590195 +0000 UTC m=+145.448030667" watchObservedRunningTime="2025-11-24 08:54:44.958531316 +0000 UTC m=+145.492971798" Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.960078 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sn69z" podStartSLOduration=125.960067151 podStartE2EDuration="2m5.960067151s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:44.958713511 +0000 UTC m=+145.493153983" watchObservedRunningTime="2025-11-24 08:54:44.960067151 +0000 UTC m=+145.494507613" Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.976888 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" event={"ID":"9eca8c62-9cfd-44c9-ad1c-5898f00e8b48","Type":"ContainerStarted","Data":"27ddbb721b9a173923a98804749ec1057e9d298ebdd4ee80894d37ee1d9a2934"} Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.980137 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.989421 4944 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-86wfc container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Nov 24 08:54:44 crc kubenswrapper[4944]: I1124 08:54:44.989491 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" podUID="9eca8c62-9cfd-44c9-ad1c-5898f00e8b48" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.004838 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.009204 4944 patch_prober.go:28] interesting pod/router-default-5444994796-bg5vl container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.009707 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bg5vl" podUID="548a2d15-cb3c-43ec-9310-9661929b878e" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.011572 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.012760 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:45.51274028 +0000 UTC m=+146.047180752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.030975 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" event={"ID":"798fde25-9b81-4d19-97e8-d9d953d58924","Type":"ContainerStarted","Data":"ffc6416ca4801dc6b136bdb29564a450aaf5086bf5e5ab76c93d2a22ce78f4ed"} Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.032210 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-bg5vl" podStartSLOduration=126.032179525 podStartE2EDuration="2m6.032179525s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:44.998373755 +0000 UTC m=+145.532814227" watchObservedRunningTime="2025-11-24 08:54:45.032179525 +0000 UTC m=+145.566619987" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.049309 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" podStartSLOduration=126.049282211 podStartE2EDuration="2m6.049282211s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:45.024321713 +0000 UTC m=+145.558762185" watchObservedRunningTime="2025-11-24 08:54:45.049282211 +0000 UTC m=+145.583722673" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.069010 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" event={"ID":"d0f91a43-15f8-425e-a5f3-9f5f15e3b506","Type":"ContainerStarted","Data":"78e88309aade1b78348cc0084dab086860e2d03d075bc6dd37d6378c59301af4"} Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.070520 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" podStartSLOduration=126.07049603 podStartE2EDuration="2m6.07049603s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:45.068818199 +0000 UTC m=+145.603258691" watchObservedRunningTime="2025-11-24 08:54:45.07049603 +0000 UTC m=+145.604936492" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.072171 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qwbqn" event={"ID":"98143ac5-35b2-4689-872f-bf65e0e402e2","Type":"ContainerStarted","Data":"45bed29735f23e6269bc8247aacfbed916b77715c3e45f2c1ef8bb5018dcfc06"} Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.079780 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" event={"ID":"2a322ced-c651-4f79-aeae-1fa31aea8fd0","Type":"ContainerStarted","Data":"3ddfb67fbd6144d3d7a0c5efbd2757443afe90c70699f8867384f517e4b6b70c"} Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.085178 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" event={"ID":"449d38e0-1f3f-46fe-a256-1dbdc041900b","Type":"ContainerStarted","Data":"fac8ea08e64bb53df07d46232b220e415fd9e013b083df9248f18822bb3e0558"} Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.100599 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-qwbqn" podStartSLOduration=7.10057128 podStartE2EDuration="7.10057128s" podCreationTimestamp="2025-11-24 08:54:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:45.089699868 +0000 UTC m=+145.624140330" watchObservedRunningTime="2025-11-24 08:54:45.10057128 +0000 UTC m=+145.635011742" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.114577 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.115916 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:45.615899624 +0000 UTC m=+146.150340086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.137275 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" podStartSLOduration=126.137245985 podStartE2EDuration="2m6.137245985s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:45.116655195 +0000 UTC m=+145.651095667" watchObservedRunningTime="2025-11-24 08:54:45.137245985 +0000 UTC m=+145.671686437" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.153006 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" event={"ID":"201c725b-a236-416d-8d93-2d0b21a103e0","Type":"ContainerStarted","Data":"35396d8591d0c31fcd2a93725f17b8e6437bbdd5dba2677b4a7c42ab87b441a3"} Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.171705 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" event={"ID":"48f168d6-d783-45a5-851a-e0090dc07f7c","Type":"ContainerStarted","Data":"31e5ba2987059ae6f87d5ba9b5e3c6ca1f750015e729b115e3640cfc0927a0fc"} Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.173960 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.194097 4944 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-d4q6h container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.194453 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" podUID="48f168d6-d783-45a5-851a-e0090dc07f7c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.201971 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" podStartSLOduration=126.20194743 podStartE2EDuration="2m6.20194743s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:45.201436455 +0000 UTC m=+145.735876927" watchObservedRunningTime="2025-11-24 08:54:45.20194743 +0000 UTC m=+145.736387892" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.213877 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" event={"ID":"f9ddea88-e412-4298-aa5c-fbd8f1938aa7","Type":"ContainerStarted","Data":"abd1c62807d792b892c9d7477af9e228b744215001f5fd9904a0b9c2f56aba87"} Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.216346 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.219024 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:45.719005265 +0000 UTC m=+146.253445727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.271813 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" podStartSLOduration=126.271784508 podStartE2EDuration="2m6.271784508s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:45.267939253 +0000 UTC m=+145.802379725" watchObservedRunningTime="2025-11-24 08:54:45.271784508 +0000 UTC m=+145.806224960" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.320687 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.320948 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:45.820903901 +0000 UTC m=+146.355344373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.321288 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.329824 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:45.829802485 +0000 UTC m=+146.364242947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.422763 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.423506 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:45.923442827 +0000 UTC m=+146.457883289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.423665 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.424400 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:45.924376184 +0000 UTC m=+146.458816646 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.524867 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.525093 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.025036083 +0000 UTC m=+146.559476545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.525455 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.525944 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.02593578 +0000 UTC m=+146.560376232 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.626301 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.627094 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.127072854 +0000 UTC m=+146.661513316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.630621 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.727891 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.728370 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.228342991 +0000 UTC m=+146.762783443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.832809 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.833538 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.333506704 +0000 UTC m=+146.867947166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.833800 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.834112 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.334100601 +0000 UTC m=+146.868541063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:45 crc kubenswrapper[4944]: I1124 08:54:45.935519 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:45 crc kubenswrapper[4944]: E1124 08:54:45.935997 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.435970647 +0000 UTC m=+146.970411109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.007035 4944 patch_prober.go:28] interesting pod/router-default-5444994796-bg5vl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:54:46 crc kubenswrapper[4944]: [-]has-synced failed: reason withheld Nov 24 08:54:46 crc kubenswrapper[4944]: [+]process-running ok Nov 24 08:54:46 crc kubenswrapper[4944]: healthz check failed Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.007123 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bg5vl" podUID="548a2d15-cb3c-43ec-9310-9661929b878e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.024341 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.024814 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.030709 4944 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-ntscp container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.14:8443/livez\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.030781 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" podUID="bb59aca8-007d-46f0-a410-75fbe41c7217" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.14:8443/livez\": dial tcp 10.217.0.14:8443: connect: connection refused" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.038359 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:46 crc kubenswrapper[4944]: E1124 08:54:46.039099 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.539079109 +0000 UTC m=+147.073519581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.140452 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:46 crc kubenswrapper[4944]: E1124 08:54:46.141246 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.641229922 +0000 UTC m=+147.175670384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.236544 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4t59q" event={"ID":"798fde25-9b81-4d19-97e8-d9d953d58924","Type":"ContainerStarted","Data":"c037bf3aaa48354a0829730537d34710263b111dd56ac40fb200b3ced627a47f"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.238616 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" event={"ID":"2a322ced-c651-4f79-aeae-1fa31aea8fd0","Type":"ContainerStarted","Data":"9128068d81f82c0370a9fa611c7cf6bd417879436d24e161ae6a3676f911afeb"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.242265 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:46 crc kubenswrapper[4944]: E1124 08:54:46.242747 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.742729876 +0000 UTC m=+147.277170338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.249668 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" event={"ID":"d0f91a43-15f8-425e-a5f3-9f5f15e3b506","Type":"ContainerStarted","Data":"fd845212fbac72b698b788d443ee5f065f53cc9a858ca5c5b53e10ca5184c9d5"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.249747 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" event={"ID":"d0f91a43-15f8-425e-a5f3-9f5f15e3b506","Type":"ContainerStarted","Data":"8c346ff22f455f0d6251a8182d88f06fba9844addf26675f75af2b01120caed2"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.253420 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" event={"ID":"46615e45-27d1-47b4-a5d7-f0cfb9f7476d","Type":"ContainerStarted","Data":"112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.254823 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.256533 4944 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-m6g58 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.256682 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" podUID="46615e45-27d1-47b4-a5d7-f0cfb9f7476d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.269952 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" event={"ID":"745f9869-752f-4731-be2e-92ee56f613b0","Type":"ContainerStarted","Data":"742fc5f7828cf62fb4e5f0ced206c10f37820b7dcf46d8a6fdaf9e6d0d474870"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.279812 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w6qvf" podStartSLOduration=127.279780203 podStartE2EDuration="2m7.279780203s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.279637499 +0000 UTC m=+146.814077971" watchObservedRunningTime="2025-11-24 08:54:46.279780203 +0000 UTC m=+146.814220685" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.326041 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v48qh" event={"ID":"449d38e0-1f3f-46fe-a256-1dbdc041900b","Type":"ContainerStarted","Data":"08e9d31f4417971078ae546fc7c6b97a63bd57a9231b2dfef5944fa7cc7d949c"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.339851 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" podStartSLOduration=127.33982677 podStartE2EDuration="2m7.33982677s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.337564793 +0000 UTC m=+146.872005255" watchObservedRunningTime="2025-11-24 08:54:46.33982677 +0000 UTC m=+146.874267242" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.353078 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:46 crc kubenswrapper[4944]: E1124 08:54:46.354681 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.854660029 +0000 UTC m=+147.389100491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.361412 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rvpws" event={"ID":"f9ddea88-e412-4298-aa5c-fbd8f1938aa7","Type":"ContainerStarted","Data":"77170c8df99bafd2c257c49b740292df9b4aefd6690ab87e3aacf708d12f51a5"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.368618 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-gvrs8" event={"ID":"1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6","Type":"ContainerStarted","Data":"8e7a1346aed18d6be0515929802bfe12e709b166a1ce1793101f9147f02c46d4"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.368897 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-gvrs8" event={"ID":"1f38bcfa-5a4e-41eb-b746-0936cbc2cbc6","Type":"ContainerStarted","Data":"9a522b738d3897605f8fd3d51d49bf198458a842eb2261c8acca56d15516ac5b"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.406977 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" event={"ID":"807b8528-f6df-46ea-80f3-564a8e0a756a","Type":"ContainerStarted","Data":"67e1e80250ee7d5e1f109e0e56bde38b75dd8eb166ac2aaf5308e25b0a4453a2"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.407023 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" event={"ID":"807b8528-f6df-46ea-80f3-564a8e0a756a","Type":"ContainerStarted","Data":"36cb2a9565ffa1fb63e1a9d46ab30ea317b2746e0fda9307a2a9209709a0de78"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.407686 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.435641 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" event={"ID":"69ba35de-8359-44fd-a9b0-e93e70f616eb","Type":"ContainerStarted","Data":"5cddb2fa65ea7ff5034884a674551a2de0d049da92083650998076b260ff6d26"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.450906 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4cm8w" podStartSLOduration=127.450882758 podStartE2EDuration="2m7.450882758s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.396322212 +0000 UTC m=+146.930762694" watchObservedRunningTime="2025-11-24 08:54:46.450882758 +0000 UTC m=+146.985323210" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.452077 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mjx2t" podStartSLOduration=127.452069292 podStartE2EDuration="2m7.452069292s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.449423754 +0000 UTC m=+146.983864216" watchObservedRunningTime="2025-11-24 08:54:46.452069292 +0000 UTC m=+146.986509754" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.458905 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:46 crc kubenswrapper[4944]: E1124 08:54:46.461825 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:46.961809111 +0000 UTC m=+147.496249573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.469376 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" event={"ID":"662092ff-6269-408c-8066-cbda91c4ecfe","Type":"ContainerStarted","Data":"6c3acc16b185e26c42dcfd0f4bdfe81af220bfd7913d9b109e2ae27bff21d78e"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.469441 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" event={"ID":"662092ff-6269-408c-8066-cbda91c4ecfe","Type":"ContainerStarted","Data":"a284edeb27c2a9a001b989a71bc25e15d2b59a6d84e771ce32082a313bf0a2d3"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.497674 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" podStartSLOduration=127.497643772 podStartE2EDuration="2m7.497643772s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.481870175 +0000 UTC m=+147.016310637" watchObservedRunningTime="2025-11-24 08:54:46.497643772 +0000 UTC m=+147.032084244" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.499138 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" event={"ID":"81e802f8-50f6-4287-9960-4722c2299155","Type":"ContainerStarted","Data":"72759c79cd1eafffa43dd31ac1bdfa2cee39811f03a2af9adb33c8e0b41de640"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.499177 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" event={"ID":"81e802f8-50f6-4287-9960-4722c2299155","Type":"ContainerStarted","Data":"8a3395e06475ad08cb611d2376ec8d8bc9a3e987400d5b78b9d20cb9078c4d48"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.515534 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-gvrs8" podStartSLOduration=8.515510211 podStartE2EDuration="8.515510211s" podCreationTimestamp="2025-11-24 08:54:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.514427598 +0000 UTC m=+147.048868060" watchObservedRunningTime="2025-11-24 08:54:46.515510211 +0000 UTC m=+147.049950673" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.530413 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" event={"ID":"f54d182a-01e5-4ac8-830a-3d73341d5099","Type":"ContainerStarted","Data":"b1347588eacfcb95fb72e4de5f01f8c944d87b0501e22a012122a0d482bd161c"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.530475 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" event={"ID":"f54d182a-01e5-4ac8-830a-3d73341d5099","Type":"ContainerStarted","Data":"082312f5ca3bdaec89b1269133ae1cd313e5f9cf337087df41ed216842b3e374"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.531599 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.540899 4944 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hwsml container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.540973 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" podUID="f54d182a-01e5-4ac8-830a-3d73341d5099" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.542587 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" event={"ID":"d4d3d329-b103-4e77-9b5b-83535f5bbc44","Type":"ContainerStarted","Data":"eee1633ae3fb879ffb9d0588acf4b698c329a6bb5e60951d82a966ede956f690"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.547258 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-fqjd6" podStartSLOduration=127.547235309 podStartE2EDuration="2m7.547235309s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.543688345 +0000 UTC m=+147.078128827" watchObservedRunningTime="2025-11-24 08:54:46.547235309 +0000 UTC m=+147.081675771" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.562864 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:46 crc kubenswrapper[4944]: E1124 08:54:46.563297 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.063262014 +0000 UTC m=+147.597702476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.565385 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" event={"ID":"201c725b-a236-416d-8d93-2d0b21a103e0","Type":"ContainerStarted","Data":"5c09caade1be245ae83bc8b7c646205c8f0a9408b3ea2018f52aa27378389243"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.570340 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7" event={"ID":"47dcb102-7868-4647-b85a-c7a6e44b66f5","Type":"ContainerStarted","Data":"fefeb7d9dd5ad7511374db831e4e01247c47adafc55e82bad6da68dd5733e8c9"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.570575 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7" event={"ID":"47dcb102-7868-4647-b85a-c7a6e44b66f5","Type":"ContainerStarted","Data":"b84c04852f0e849e4539b96b85d2829ddfc18bb417490ac2ab775e3be68149e6"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.574039 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" event={"ID":"9448bfcf-6ec7-41ed-a6b5-f3243dec96af","Type":"ContainerStarted","Data":"c89b7e0041b410fca1eda3b4f7dcf44ab03b6e22b6691e91808f51912e791b54"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.575035 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" event={"ID":"0768e45f-6339-47f6-ba67-ffe401d2a502","Type":"ContainerStarted","Data":"f8c7fff3a079662f189454565ca21ea4eb2ffa8ce1ccecdd2eb63949e41cae04"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.575168 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" event={"ID":"0768e45f-6339-47f6-ba67-ffe401d2a502","Type":"ContainerStarted","Data":"810433f3eb6c738001917a77edf10fdfa8279acbef5658e4ffe227c8799c3d21"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.576829 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv" event={"ID":"22e1dda2-1c29-4f22-97c3-71c6aa588b47","Type":"ContainerStarted","Data":"1b591ada881607f647ac34861c388d7ea5ca3daa66dfab5ccc7d12515dd0c5b5"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.576924 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv" event={"ID":"22e1dda2-1c29-4f22-97c3-71c6aa588b47","Type":"ContainerStarted","Data":"430210cb903cd34613f7305fbcf4f12b576d76fc2ff881091e734b8629b30046"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.597314 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8m4x8" podStartSLOduration=127.59728019 podStartE2EDuration="2m7.59728019s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.586308236 +0000 UTC m=+147.120748698" watchObservedRunningTime="2025-11-24 08:54:46.59728019 +0000 UTC m=+147.131720652" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.645582 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" event={"ID":"bb59aca8-007d-46f0-a410-75fbe41c7217","Type":"ContainerStarted","Data":"73f3bebf9b6b34636d1758d24b630f0b5e029501833f53afd8e4af9571db6b29"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.650742 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jz9hv" podStartSLOduration=127.650718112 podStartE2EDuration="2m7.650718112s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.646101236 +0000 UTC m=+147.180541698" watchObservedRunningTime="2025-11-24 08:54:46.650718112 +0000 UTC m=+147.185158574" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.651925 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-68hsd" event={"ID":"5fcd0548-fb06-43f1-9328-a0f58b7db2da","Type":"ContainerStarted","Data":"ef3c83a0fde05a382a6b1921b2bf74b527a3be4d92e24da8565cdf44f2c4b799"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.651972 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-68hsd" event={"ID":"5fcd0548-fb06-43f1-9328-a0f58b7db2da","Type":"ContainerStarted","Data":"aa91e5f0be8df7607c9fefb1855eddcf4c334ff4d6f2d2dcb0725e0854410df0"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.652410 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.658167 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" event={"ID":"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0","Type":"ContainerStarted","Data":"c39fc89d6a5d00ff2ac7c7b970c5133d0c2e0d910f51385faec8653104f9f938"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.658198 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" event={"ID":"8868f76b-cfe5-4fb4-8866-e0b5ee27d3f0","Type":"ContainerStarted","Data":"0d8a16a8d049eac1eba65320bd381d10c212f34a47c491c7f97a1bf007f14839"} Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.658642 4944 patch_prober.go:28] interesting pod/downloads-7954f5f757-9tglc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.658673 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9tglc" podUID="bbaf6605-39cb-41a0-900a-b2a87a00751e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.666942 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.675242 4944 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-x4nbt container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.675336 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" podUID="3dc08e9e-07bc-4330-8caa-d73b9ac7faf4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Nov 24 08:54:46 crc kubenswrapper[4944]: E1124 08:54:46.675732 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.175711362 +0000 UTC m=+147.710152024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.714661 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" podStartSLOduration=127.714641214 podStartE2EDuration="2m7.714641214s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.705141563 +0000 UTC m=+147.239582025" watchObservedRunningTime="2025-11-24 08:54:46.714641214 +0000 UTC m=+147.249081666" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.733613 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d4q6h" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.740328 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-bkwvc" podStartSLOduration=127.740305964 podStartE2EDuration="2m7.740305964s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.739413518 +0000 UTC m=+147.273854000" watchObservedRunningTime="2025-11-24 08:54:46.740305964 +0000 UTC m=+147.274746426" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.768345 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:46 crc kubenswrapper[4944]: E1124 08:54:46.770775 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.270717804 +0000 UTC m=+147.805158266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.815238 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7" podStartSLOduration=127.815213952 podStartE2EDuration="2m7.815213952s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.811431029 +0000 UTC m=+147.345871501" watchObservedRunningTime="2025-11-24 08:54:46.815213952 +0000 UTC m=+147.349654414" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.815332 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" podStartSLOduration=127.815328605 podStartE2EDuration="2m7.815328605s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.78947797 +0000 UTC m=+147.323918442" watchObservedRunningTime="2025-11-24 08:54:46.815328605 +0000 UTC m=+147.349769067" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.851331 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-hf8bj" podStartSLOduration=127.85130351 podStartE2EDuration="2m7.85130351s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.848545738 +0000 UTC m=+147.382986200" watchObservedRunningTime="2025-11-24 08:54:46.85130351 +0000 UTC m=+147.385743972" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.877201 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:46 crc kubenswrapper[4944]: E1124 08:54:46.877668 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.377647519 +0000 UTC m=+147.912087971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.919441 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-68hsd" podStartSLOduration=8.919416195 podStartE2EDuration="8.919416195s" podCreationTimestamp="2025-11-24 08:54:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.890040806 +0000 UTC m=+147.424481278" watchObservedRunningTime="2025-11-24 08:54:46.919416195 +0000 UTC m=+147.453856657" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.972335 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-86wjt" podStartSLOduration=127.972311861 podStartE2EDuration="2m7.972311861s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:46.97026773 +0000 UTC m=+147.504708192" watchObservedRunningTime="2025-11-24 08:54:46.972311861 +0000 UTC m=+147.506752323" Nov 24 08:54:46 crc kubenswrapper[4944]: I1124 08:54:46.978974 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:46 crc kubenswrapper[4944]: E1124 08:54:46.979507 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.479486204 +0000 UTC m=+148.013926666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.002831 4944 patch_prober.go:28] interesting pod/router-default-5444994796-bg5vl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:54:47 crc kubenswrapper[4944]: [-]has-synced failed: reason withheld Nov 24 08:54:47 crc kubenswrapper[4944]: [+]process-running ok Nov 24 08:54:47 crc kubenswrapper[4944]: healthz check failed Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.002899 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bg5vl" podUID="548a2d15-cb3c-43ec-9310-9661929b878e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.077469 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-86wfc" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.081358 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:47 crc kubenswrapper[4944]: E1124 08:54:47.081764 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.58174894 +0000 UTC m=+148.116189412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.182697 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.183648 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.183858 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:47 crc kubenswrapper[4944]: E1124 08:54:47.185136 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.685113429 +0000 UTC m=+148.219553901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.188804 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.201160 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.201450 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.284938 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.285022 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.285116 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:47 crc kubenswrapper[4944]: E1124 08:54:47.290087 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.790063146 +0000 UTC m=+148.324503608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.291090 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.301574 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.385967 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:47 crc kubenswrapper[4944]: E1124 08:54:47.386823 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.886802949 +0000 UTC m=+148.421243421 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.490079 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:47 crc kubenswrapper[4944]: E1124 08:54:47.490555 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:47.99053879 +0000 UTC m=+148.524979252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.509409 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.515363 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.592818 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:47 crc kubenswrapper[4944]: E1124 08:54:47.593203 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:48.093183388 +0000 UTC m=+148.627623850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.684090 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-b7gc7" event={"ID":"47dcb102-7868-4647-b85a-c7a6e44b66f5","Type":"ContainerStarted","Data":"95e7e931ba1089bbf444448209dac459a65bbb6df4ab0799dd8e1e0248cb66bf"} Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.693902 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:47 crc kubenswrapper[4944]: E1124 08:54:47.694388 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:48.194375734 +0000 UTC m=+148.728816196 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.697455 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" event={"ID":"9448bfcf-6ec7-41ed-a6b5-f3243dec96af","Type":"ContainerStarted","Data":"13febb31cd30c96a794666e7ea7af59e91750f9f12a6cbef37517916469a3679"} Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.710331 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"cf989a12487cb4f66bc742b22246b5e9fd09adf6bb56d33b80305c6f2b51c8f7"} Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.728031 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" event={"ID":"69ba35de-8359-44fd-a9b0-e93e70f616eb","Type":"ContainerStarted","Data":"82cf31d9fc9964c850bfa154a4051057f99d624004e5a29bff8e550067aec7bd"} Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.729559 4944 patch_prober.go:28] interesting pod/downloads-7954f5f757-9tglc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.729608 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9tglc" podUID="bbaf6605-39cb-41a0-900a-b2a87a00751e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.731505 4944 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-m6g58 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.731558 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" podUID="46615e45-27d1-47b4-a5d7-f0cfb9f7476d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.757481 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hwsml" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.794981 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:47 crc kubenswrapper[4944]: E1124 08:54:47.796884 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:48.296839486 +0000 UTC m=+148.831279968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.808762 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" podStartSLOduration=128.808740679 podStartE2EDuration="2m8.808740679s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:47.805767151 +0000 UTC m=+148.340207613" watchObservedRunningTime="2025-11-24 08:54:47.808740679 +0000 UTC m=+148.343181141" Nov 24 08:54:47 crc kubenswrapper[4944]: I1124 08:54:47.900320 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:47 crc kubenswrapper[4944]: E1124 08:54:47.905777 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:48.40575888 +0000 UTC m=+148.940199552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.005542 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.005934 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:48.505915114 +0000 UTC m=+149.040355586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.014326 4944 patch_prober.go:28] interesting pod/router-default-5444994796-bg5vl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:54:48 crc kubenswrapper[4944]: [-]has-synced failed: reason withheld Nov 24 08:54:48 crc kubenswrapper[4944]: [+]process-running ok Nov 24 08:54:48 crc kubenswrapper[4944]: healthz check failed Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.014401 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bg5vl" podUID="548a2d15-cb3c-43ec-9310-9661929b878e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.108606 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.109096 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:48.609080318 +0000 UTC m=+149.143520790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.209813 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.210071 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:48.710012135 +0000 UTC m=+149.244452607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.210129 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.210518 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:48.7105033 +0000 UTC m=+149.244943762 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.311111 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.311527 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:48.811495629 +0000 UTC m=+149.345936091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.311870 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.316403 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:48.816366463 +0000 UTC m=+149.350806925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.337090 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x4nbt" Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.413001 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.413498 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:48.913474468 +0000 UTC m=+149.447914930 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.514602 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.515276 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.01525643 +0000 UTC m=+149.549696882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.615737 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.616543 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.116511638 +0000 UTC m=+149.650952100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.717815 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.718426 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.218402684 +0000 UTC m=+149.752843326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.733190 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"20380dd1479f8f9db75da46c1158e8ed8eebe553b0d67624a979f55d8e36c3e5"} Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.733241 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"f4bf1c77f7df03650c5a7358741815c00a6a8a33c45167c03748aee81057eb33"} Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.735994 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"571092f46e3f209c8db4d7b666821eae796b15b95f835bd5641772a7a5e6c9a1"} Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.742294 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a6cb11fb70c6f9015f7b227ccadcac70bcb91df5fdefe5d5aca66befd52a69e6"} Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.742353 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f0587729608a8d48cd282c667f4a66f4b1bdc29fdc04e9f5fb9592b2a1cc8e25"} Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.752205 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.818563 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.823172 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.323140313 +0000 UTC m=+149.857580775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.824683 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.828808 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.328794371 +0000 UTC m=+149.863234833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:48 crc kubenswrapper[4944]: I1124 08:54:48.926892 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:48 crc kubenswrapper[4944]: E1124 08:54:48.927281 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.427263425 +0000 UTC m=+149.961703887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.005056 4944 patch_prober.go:28] interesting pod/router-default-5444994796-bg5vl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:54:49 crc kubenswrapper[4944]: [-]has-synced failed: reason withheld Nov 24 08:54:49 crc kubenswrapper[4944]: [+]process-running ok Nov 24 08:54:49 crc kubenswrapper[4944]: healthz check failed Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.005132 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bg5vl" podUID="548a2d15-cb3c-43ec-9310-9661929b878e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.028767 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.029152 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.529140751 +0000 UTC m=+150.063581213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.130181 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.130387 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.630353947 +0000 UTC m=+150.164794419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.130523 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.130893 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.630882392 +0000 UTC m=+150.165322864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.232036 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.232168 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.732128839 +0000 UTC m=+150.266569301 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.232324 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.232692 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.732684066 +0000 UTC m=+150.267124528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.338407 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.339131 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.839113786 +0000 UTC m=+150.373554248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.339174 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.339565 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.839542249 +0000 UTC m=+150.373982701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.429632 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8cs79"] Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.430616 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.435915 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.440223 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.440688 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:49.940666062 +0000 UTC m=+150.475106524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.443316 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8cs79"] Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.500279 4944 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.542335 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxnxt\" (UniqueName: \"kubernetes.io/projected/c8249efc-c3ac-4527-8394-e6e0aff9d457-kube-api-access-kxnxt\") pod \"community-operators-8cs79\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.542422 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-catalog-content\") pod \"community-operators-8cs79\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.542573 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-utilities\") pod \"community-operators-8cs79\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.542740 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.543271 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:50.043251238 +0000 UTC m=+150.577691700 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.620667 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sphtd"] Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.621945 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.625267 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.629682 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sphtd"] Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.643508 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.644000 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxnxt\" (UniqueName: \"kubernetes.io/projected/c8249efc-c3ac-4527-8394-e6e0aff9d457-kube-api-access-kxnxt\") pod \"community-operators-8cs79\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.644117 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-catalog-content\") pod \"community-operators-8cs79\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.644168 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-utilities\") pod \"community-operators-8cs79\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.644681 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-utilities\") pod \"community-operators-8cs79\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.644766 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:50.144750063 +0000 UTC m=+150.679190525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.645657 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-catalog-content\") pod \"community-operators-8cs79\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.666739 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxnxt\" (UniqueName: \"kubernetes.io/projected/c8249efc-c3ac-4527-8394-e6e0aff9d457-kube-api-access-kxnxt\") pod \"community-operators-8cs79\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.745357 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.745430 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlj4f\" (UniqueName: \"kubernetes.io/projected/9f532fb4-a915-4cf7-a64a-887eadfb1710-kube-api-access-hlj4f\") pod \"certified-operators-sphtd\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.745463 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-utilities\") pod \"certified-operators-sphtd\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.745493 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-catalog-content\") pod \"certified-operators-sphtd\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.746130 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:50.246117833 +0000 UTC m=+150.780558295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.752819 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" event={"ID":"9448bfcf-6ec7-41ed-a6b5-f3243dec96af","Type":"ContainerStarted","Data":"dac0a37f016a756f5987e15b2558e78bbcc77fba314f85cbc548da98f17be8be"} Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.752870 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" event={"ID":"9448bfcf-6ec7-41ed-a6b5-f3243dec96af","Type":"ContainerStarted","Data":"6f35fb23ffa16f52f24875d22e40ab6af184742209a69d680232dfee4f67ff11"} Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.752880 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" event={"ID":"9448bfcf-6ec7-41ed-a6b5-f3243dec96af","Type":"ContainerStarted","Data":"02dcb187126334e6ce408fd7d6beac06b44bbf774edde559e3dc0509e913a148"} Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.753898 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.755829 4944 generic.go:334] "Generic (PLEG): container finished" podID="0768e45f-6339-47f6-ba67-ffe401d2a502" containerID="f8c7fff3a079662f189454565ca21ea4eb2ffa8ce1ccecdd2eb63949e41cae04" exitCode=0 Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.756366 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" event={"ID":"0768e45f-6339-47f6-ba67-ffe401d2a502","Type":"ContainerDied","Data":"f8c7fff3a079662f189454565ca21ea4eb2ffa8ce1ccecdd2eb63949e41cae04"} Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.780105 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-jtvv4" podStartSLOduration=11.780082709 podStartE2EDuration="11.780082709s" podCreationTimestamp="2025-11-24 08:54:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:49.778314186 +0000 UTC m=+150.312754638" watchObservedRunningTime="2025-11-24 08:54:49.780082709 +0000 UTC m=+150.314523181" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.832026 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vsqrm"] Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.833713 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.846492 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.846628 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:50.346606427 +0000 UTC m=+150.881046899 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.846794 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlj4f\" (UniqueName: \"kubernetes.io/projected/9f532fb4-a915-4cf7-a64a-887eadfb1710-kube-api-access-hlj4f\") pod \"certified-operators-sphtd\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.846861 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-utilities\") pod \"certified-operators-sphtd\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.846910 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-catalog-content\") pod \"certified-operators-sphtd\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.847021 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.847353 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:50.347343279 +0000 UTC m=+150.881783741 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.847800 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-utilities\") pod \"certified-operators-sphtd\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.848140 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-catalog-content\") pod \"certified-operators-sphtd\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.855246 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vsqrm"] Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.890212 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlj4f\" (UniqueName: \"kubernetes.io/projected/9f532fb4-a915-4cf7-a64a-887eadfb1710-kube-api-access-hlj4f\") pod \"certified-operators-sphtd\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.953403 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.956879 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.957107 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-utilities\") pod \"community-operators-vsqrm\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.957280 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-catalog-content\") pod \"community-operators-vsqrm\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:54:49 crc kubenswrapper[4944]: I1124 08:54:49.957306 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndb9f\" (UniqueName: \"kubernetes.io/projected/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-kube-api-access-ndb9f\") pod \"community-operators-vsqrm\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:54:49 crc kubenswrapper[4944]: E1124 08:54:49.957417 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:50.457401826 +0000 UTC m=+150.991842278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.002417 4944 patch_prober.go:28] interesting pod/router-default-5444994796-bg5vl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:54:50 crc kubenswrapper[4944]: [-]has-synced failed: reason withheld Nov 24 08:54:50 crc kubenswrapper[4944]: [+]process-running ok Nov 24 08:54:50 crc kubenswrapper[4944]: healthz check failed Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.002484 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bg5vl" podUID="548a2d15-cb3c-43ec-9310-9661929b878e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.045194 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9cxw2"] Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.047727 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.058691 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9cxw2"] Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.071350 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.071529 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-catalog-content\") pod \"community-operators-vsqrm\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.071556 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndb9f\" (UniqueName: \"kubernetes.io/projected/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-kube-api-access-ndb9f\") pod \"community-operators-vsqrm\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.071616 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-utilities\") pod \"community-operators-vsqrm\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:54:50 crc kubenswrapper[4944]: E1124 08:54:50.071855 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:50.571839754 +0000 UTC m=+151.106280216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.072117 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-utilities\") pod \"community-operators-vsqrm\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.072324 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-catalog-content\") pod \"community-operators-vsqrm\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.109200 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndb9f\" (UniqueName: \"kubernetes.io/projected/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-kube-api-access-ndb9f\") pod \"community-operators-vsqrm\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.140952 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8cs79"] Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.152860 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.172453 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:50 crc kubenswrapper[4944]: E1124 08:54:50.172690 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:50.672654398 +0000 UTC m=+151.207094870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.172904 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-utilities\") pod \"certified-operators-9cxw2\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.172969 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxhmm\" (UniqueName: \"kubernetes.io/projected/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-kube-api-access-jxhmm\") pod \"certified-operators-9cxw2\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.173019 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.173059 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-catalog-content\") pod \"certified-operators-9cxw2\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:54:50 crc kubenswrapper[4944]: E1124 08:54:50.173368 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:50.673357119 +0000 UTC m=+151.207797771 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.274105 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.274371 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxhmm\" (UniqueName: \"kubernetes.io/projected/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-kube-api-access-jxhmm\") pod \"certified-operators-9cxw2\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.274460 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-catalog-content\") pod \"certified-operators-9cxw2\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.274490 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-utilities\") pod \"certified-operators-9cxw2\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.274913 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-utilities\") pod \"certified-operators-9cxw2\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:54:50 crc kubenswrapper[4944]: E1124 08:54:50.275004 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:54:50.774985007 +0000 UTC m=+151.309425479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.275589 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-catalog-content\") pod \"certified-operators-9cxw2\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.306669 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxhmm\" (UniqueName: \"kubernetes.io/projected/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-kube-api-access-jxhmm\") pod \"certified-operators-9cxw2\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.314193 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sphtd"] Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.378750 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:50 crc kubenswrapper[4944]: E1124 08:54:50.379220 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:54:50.879206592 +0000 UTC m=+151.413647054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-v5288" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.426486 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.441396 4944 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T08:54:49.500315077Z","Handler":null,"Name":""} Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.459006 4944 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.459058 4944 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.476386 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vsqrm"] Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.479476 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.486509 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.581446 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.584029 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.584079 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.615703 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-v5288\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.654548 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9cxw2"] Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.707130 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.734392 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.734534 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.738217 4944 patch_prober.go:28] interesting pod/console-f9d7485db-s9r2p container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.738272 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-s9r2p" podUID="b355df00-f706-42c5-b16f-8c237a5048d8" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.777828 4944 generic.go:334] "Generic (PLEG): container finished" podID="9f532fb4-a915-4cf7-a64a-887eadfb1710" containerID="5615ab50d9653d288de46ce0dffbb16d51a4f75ec6475a0c60212214fc0c6224" exitCode=0 Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.778203 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sphtd" event={"ID":"9f532fb4-a915-4cf7-a64a-887eadfb1710","Type":"ContainerDied","Data":"5615ab50d9653d288de46ce0dffbb16d51a4f75ec6475a0c60212214fc0c6224"} Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.778270 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sphtd" event={"ID":"9f532fb4-a915-4cf7-a64a-887eadfb1710","Type":"ContainerStarted","Data":"30d9b2775ddd90a1aec2a749f5f5fffb9ac403b0c9cbfdeba22eda7a27bbd80e"} Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.780468 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.786116 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxw2" event={"ID":"a6587b50-bc2c-407f-b0f1-f22b6ea202cc","Type":"ContainerStarted","Data":"c396ee2079dd470d2cd231c3dc183261c6c36467b0e807b9783c27770276463c"} Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.790953 4944 generic.go:334] "Generic (PLEG): container finished" podID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" containerID="2f5c19aa2d44475dbec454564398e9bceeceb6439b4f31da4a5b53386cc28ec4" exitCode=0 Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.791122 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vsqrm" event={"ID":"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab","Type":"ContainerDied","Data":"2f5c19aa2d44475dbec454564398e9bceeceb6439b4f31da4a5b53386cc28ec4"} Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.791163 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vsqrm" event={"ID":"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab","Type":"ContainerStarted","Data":"81bb06da389a25a2402232d390c51cdec29e7cdf5a908349aa264cad993bd339"} Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.800323 4944 generic.go:334] "Generic (PLEG): container finished" podID="c8249efc-c3ac-4527-8394-e6e0aff9d457" containerID="beaf3a9957410846a72137dc26b28e5cae51d9ed97c7a96cb087a05358fdff02" exitCode=0 Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.800868 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8cs79" event={"ID":"c8249efc-c3ac-4527-8394-e6e0aff9d457","Type":"ContainerDied","Data":"beaf3a9957410846a72137dc26b28e5cae51d9ed97c7a96cb087a05358fdff02"} Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.800932 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8cs79" event={"ID":"c8249efc-c3ac-4527-8394-e6e0aff9d457","Type":"ContainerStarted","Data":"e316def9707a1c3e9f2032c1f56b141c0955ac2a85c06b800e6ddec1da3c1079"} Nov 24 08:54:50 crc kubenswrapper[4944]: I1124 08:54:50.961577 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v5288"] Nov 24 08:54:50 crc kubenswrapper[4944]: W1124 08:54:50.965952 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17c159e8_7de8_4716_b4cd_7f28a257a2bb.slice/crio-96bda08e2027b94d28c583ee2b21129486167239153126cc50297909a911e15b WatchSource:0}: Error finding container 96bda08e2027b94d28c583ee2b21129486167239153126cc50297909a911e15b: Status 404 returned error can't find the container with id 96bda08e2027b94d28c583ee2b21129486167239153126cc50297909a911e15b Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.003170 4944 patch_prober.go:28] interesting pod/router-default-5444994796-bg5vl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:54:51 crc kubenswrapper[4944]: [-]has-synced failed: reason withheld Nov 24 08:54:51 crc kubenswrapper[4944]: [+]process-running ok Nov 24 08:54:51 crc kubenswrapper[4944]: healthz check failed Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.003271 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bg5vl" podUID="548a2d15-cb3c-43ec-9310-9661929b878e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.034236 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.044491 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ntscp" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.103664 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.144175 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 08:54:51 crc kubenswrapper[4944]: E1124 08:54:51.144445 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0768e45f-6339-47f6-ba67-ffe401d2a502" containerName="collect-profiles" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.144461 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0768e45f-6339-47f6-ba67-ffe401d2a502" containerName="collect-profiles" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.144606 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0768e45f-6339-47f6-ba67-ffe401d2a502" containerName="collect-profiles" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.145079 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.149456 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.149672 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.175652 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.194657 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6tl5\" (UniqueName: \"kubernetes.io/projected/0768e45f-6339-47f6-ba67-ffe401d2a502-kube-api-access-c6tl5\") pod \"0768e45f-6339-47f6-ba67-ffe401d2a502\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.195092 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0768e45f-6339-47f6-ba67-ffe401d2a502-config-volume\") pod \"0768e45f-6339-47f6-ba67-ffe401d2a502\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.195153 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0768e45f-6339-47f6-ba67-ffe401d2a502-secret-volume\") pod \"0768e45f-6339-47f6-ba67-ffe401d2a502\" (UID: \"0768e45f-6339-47f6-ba67-ffe401d2a502\") " Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.197231 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0768e45f-6339-47f6-ba67-ffe401d2a502-config-volume" (OuterVolumeSpecName: "config-volume") pod "0768e45f-6339-47f6-ba67-ffe401d2a502" (UID: "0768e45f-6339-47f6-ba67-ffe401d2a502"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.218206 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0768e45f-6339-47f6-ba67-ffe401d2a502-kube-api-access-c6tl5" (OuterVolumeSpecName: "kube-api-access-c6tl5") pod "0768e45f-6339-47f6-ba67-ffe401d2a502" (UID: "0768e45f-6339-47f6-ba67-ffe401d2a502"). InnerVolumeSpecName "kube-api-access-c6tl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.218477 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0768e45f-6339-47f6-ba67-ffe401d2a502-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0768e45f-6339-47f6-ba67-ffe401d2a502" (UID: "0768e45f-6339-47f6-ba67-ffe401d2a502"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.296539 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38028fcd-3200-4382-967a-02637d30f33a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"38028fcd-3200-4382-967a-02637d30f33a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.296668 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38028fcd-3200-4382-967a-02637d30f33a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"38028fcd-3200-4382-967a-02637d30f33a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.296767 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0768e45f-6339-47f6-ba67-ffe401d2a502-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.296787 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0768e45f-6339-47f6-ba67-ffe401d2a502-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.296802 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6tl5\" (UniqueName: \"kubernetes.io/projected/0768e45f-6339-47f6-ba67-ffe401d2a502-kube-api-access-c6tl5\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.397613 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38028fcd-3200-4382-967a-02637d30f33a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"38028fcd-3200-4382-967a-02637d30f33a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.397777 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38028fcd-3200-4382-967a-02637d30f33a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"38028fcd-3200-4382-967a-02637d30f33a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.397719 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38028fcd-3200-4382-967a-02637d30f33a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"38028fcd-3200-4382-967a-02637d30f33a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.408305 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.409409 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.421461 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.426287 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6dlz8"] Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.427529 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.430407 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.432142 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38028fcd-3200-4382-967a-02637d30f33a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"38028fcd-3200-4382-967a-02637d30f33a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.441088 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6dlz8"] Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.470521 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.573848 4944 patch_prober.go:28] interesting pod/downloads-7954f5f757-9tglc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.576210 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9tglc" podUID="bbaf6605-39cb-41a0-900a-b2a87a00751e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.575091 4944 patch_prober.go:28] interesting pod/downloads-7954f5f757-9tglc container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.576622 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9tglc" podUID="bbaf6605-39cb-41a0-900a-b2a87a00751e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.600213 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5dwm\" (UniqueName: \"kubernetes.io/projected/32b94450-1b99-4fc9-aac8-658628a2195f-kube-api-access-l5dwm\") pod \"redhat-marketplace-6dlz8\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.600307 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-utilities\") pod \"redhat-marketplace-6dlz8\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.600340 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-catalog-content\") pod \"redhat-marketplace-6dlz8\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.668087 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.701392 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5dwm\" (UniqueName: \"kubernetes.io/projected/32b94450-1b99-4fc9-aac8-658628a2195f-kube-api-access-l5dwm\") pod \"redhat-marketplace-6dlz8\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.701552 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-utilities\") pod \"redhat-marketplace-6dlz8\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.701599 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-catalog-content\") pod \"redhat-marketplace-6dlz8\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.702373 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-catalog-content\") pod \"redhat-marketplace-6dlz8\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.702475 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-utilities\") pod \"redhat-marketplace-6dlz8\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.722390 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5dwm\" (UniqueName: \"kubernetes.io/projected/32b94450-1b99-4fc9-aac8-658628a2195f-kube-api-access-l5dwm\") pod \"redhat-marketplace-6dlz8\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.762425 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.814712 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"38028fcd-3200-4382-967a-02637d30f33a","Type":"ContainerStarted","Data":"bc92b10816338e17bb60f27e416914feb3281596c66570af41f6451fa865090e"} Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.823690 4944 generic.go:334] "Generic (PLEG): container finished" podID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" containerID="e9175e625095da4269400c2f8310ef0e9dce4efceb2006987379432ef18e3473" exitCode=0 Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.823815 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxw2" event={"ID":"a6587b50-bc2c-407f-b0f1-f22b6ea202cc","Type":"ContainerDied","Data":"e9175e625095da4269400c2f8310ef0e9dce4efceb2006987379432ef18e3473"} Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.824626 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4s66h"] Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.825868 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.833420 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" event={"ID":"17c159e8-7de8-4716-b4cd-7f28a257a2bb","Type":"ContainerStarted","Data":"4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f"} Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.833486 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" event={"ID":"17c159e8-7de8-4716-b4cd-7f28a257a2bb","Type":"ContainerStarted","Data":"96bda08e2027b94d28c583ee2b21129486167239153126cc50297909a911e15b"} Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.834371 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.860238 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s66h"] Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.886208 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" event={"ID":"0768e45f-6339-47f6-ba67-ffe401d2a502","Type":"ContainerDied","Data":"810433f3eb6c738001917a77edf10fdfa8279acbef5658e4ffe227c8799c3d21"} Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.886752 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="810433f3eb6c738001917a77edf10fdfa8279acbef5658e4ffe227c8799c3d21" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.886431 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.901059 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-tm2ss" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.903812 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-catalog-content\") pod \"redhat-marketplace-4s66h\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.903950 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b68fl\" (UniqueName: \"kubernetes.io/projected/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-kube-api-access-b68fl\") pod \"redhat-marketplace-4s66h\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.904071 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-utilities\") pod \"redhat-marketplace-4s66h\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.913452 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" podStartSLOduration=132.913425372 podStartE2EDuration="2m12.913425372s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:51.912664891 +0000 UTC m=+152.447105353" watchObservedRunningTime="2025-11-24 08:54:51.913425372 +0000 UTC m=+152.447865844" Nov 24 08:54:51 crc kubenswrapper[4944]: I1124 08:54:51.999449 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.004686 4944 patch_prober.go:28] interesting pod/router-default-5444994796-bg5vl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:54:52 crc kubenswrapper[4944]: [-]has-synced failed: reason withheld Nov 24 08:54:52 crc kubenswrapper[4944]: [+]process-running ok Nov 24 08:54:52 crc kubenswrapper[4944]: healthz check failed Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.004751 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bg5vl" podUID="548a2d15-cb3c-43ec-9310-9661929b878e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.005636 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-catalog-content\") pod \"redhat-marketplace-4s66h\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.005815 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b68fl\" (UniqueName: \"kubernetes.io/projected/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-kube-api-access-b68fl\") pod \"redhat-marketplace-4s66h\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.005847 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-utilities\") pod \"redhat-marketplace-4s66h\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.006878 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-utilities\") pod \"redhat-marketplace-4s66h\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.008011 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-catalog-content\") pod \"redhat-marketplace-4s66h\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.042896 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b68fl\" (UniqueName: \"kubernetes.io/projected/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-kube-api-access-b68fl\") pod \"redhat-marketplace-4s66h\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.156889 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.294802 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.296660 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6dlz8"] Nov 24 08:54:52 crc kubenswrapper[4944]: W1124 08:54:52.317296 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32b94450_1b99_4fc9_aac8_658628a2195f.slice/crio-f20602614f18845f7da5fdbf6e3ff1c37310e2ce58d6484cdd31ac5ccd019213 WatchSource:0}: Error finding container f20602614f18845f7da5fdbf6e3ff1c37310e2ce58d6484cdd31ac5ccd019213: Status 404 returned error can't find the container with id f20602614f18845f7da5fdbf6e3ff1c37310e2ce58d6484cdd31ac5ccd019213 Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.474481 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s66h"] Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.629925 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9q6bg"] Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.638361 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.640351 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9q6bg"] Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.641435 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.731827 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txvbs\" (UniqueName: \"kubernetes.io/projected/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-kube-api-access-txvbs\") pod \"redhat-operators-9q6bg\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.731921 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-utilities\") pod \"redhat-operators-9q6bg\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.731953 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-catalog-content\") pod \"redhat-operators-9q6bg\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.832736 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txvbs\" (UniqueName: \"kubernetes.io/projected/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-kube-api-access-txvbs\") pod \"redhat-operators-9q6bg\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.832810 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-utilities\") pod \"redhat-operators-9q6bg\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.832831 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-catalog-content\") pod \"redhat-operators-9q6bg\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.833357 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-catalog-content\") pod \"redhat-operators-9q6bg\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.833845 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-utilities\") pod \"redhat-operators-9q6bg\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.862340 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txvbs\" (UniqueName: \"kubernetes.io/projected/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-kube-api-access-txvbs\") pod \"redhat-operators-9q6bg\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.904444 4944 generic.go:334] "Generic (PLEG): container finished" podID="38028fcd-3200-4382-967a-02637d30f33a" containerID="5b24c4528f64a7a19fe2bbefa9db4e6adf69381c8e4d4bdae9c8f3875a570fc6" exitCode=0 Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.904539 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"38028fcd-3200-4382-967a-02637d30f33a","Type":"ContainerDied","Data":"5b24c4528f64a7a19fe2bbefa9db4e6adf69381c8e4d4bdae9c8f3875a570fc6"} Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.907287 4944 generic.go:334] "Generic (PLEG): container finished" podID="32b94450-1b99-4fc9-aac8-658628a2195f" containerID="8655e4f1a7c5135556bf82b8285747f1c58c7e3c90c8d25f69c2205e80f2c782" exitCode=0 Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.907367 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6dlz8" event={"ID":"32b94450-1b99-4fc9-aac8-658628a2195f","Type":"ContainerDied","Data":"8655e4f1a7c5135556bf82b8285747f1c58c7e3c90c8d25f69c2205e80f2c782"} Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.907391 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6dlz8" event={"ID":"32b94450-1b99-4fc9-aac8-658628a2195f","Type":"ContainerStarted","Data":"f20602614f18845f7da5fdbf6e3ff1c37310e2ce58d6484cdd31ac5ccd019213"} Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.914528 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s66h" event={"ID":"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf","Type":"ContainerStarted","Data":"bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629"} Nov 24 08:54:52 crc kubenswrapper[4944]: I1124 08:54:52.914575 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s66h" event={"ID":"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf","Type":"ContainerStarted","Data":"1bfc78bbd7b596e3664d97e815b0f853e8be6f7aec57808ae3b524fc8129f52f"} Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.009797 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.013685 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-bg5vl" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.032479 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pcq2j"] Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.033693 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.052635 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pcq2j"] Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.078511 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.137604 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-utilities\") pod \"redhat-operators-pcq2j\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.137683 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-catalog-content\") pod \"redhat-operators-pcq2j\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.137734 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xntk6\" (UniqueName: \"kubernetes.io/projected/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-kube-api-access-xntk6\") pod \"redhat-operators-pcq2j\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.240234 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-catalog-content\") pod \"redhat-operators-pcq2j\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.240941 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-catalog-content\") pod \"redhat-operators-pcq2j\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.240981 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xntk6\" (UniqueName: \"kubernetes.io/projected/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-kube-api-access-xntk6\") pod \"redhat-operators-pcq2j\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.241151 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-utilities\") pod \"redhat-operators-pcq2j\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.243092 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-utilities\") pod \"redhat-operators-pcq2j\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.260690 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xntk6\" (UniqueName: \"kubernetes.io/projected/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-kube-api-access-xntk6\") pod \"redhat-operators-pcq2j\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.391333 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.429683 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9q6bg"] Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.556119 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.556537 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.792684 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pcq2j"] Nov 24 08:54:53 crc kubenswrapper[4944]: W1124 08:54:53.799274 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb7721e8_723b_435f_bb1c_e4b0fcf3372b.slice/crio-a45d285255c191fb0afde701452164214a94ce3c5758d26040644d9adeda753e WatchSource:0}: Error finding container a45d285255c191fb0afde701452164214a94ce3c5758d26040644d9adeda753e: Status 404 returned error can't find the container with id a45d285255c191fb0afde701452164214a94ce3c5758d26040644d9adeda753e Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.942347 4944 generic.go:334] "Generic (PLEG): container finished" podID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerID="6ac9bb1fd81282b0a1c6bdb0d059a6a3560d3710c8a1d00a1a936867d8053db9" exitCode=0 Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.942436 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6bg" event={"ID":"6a34571c-5dc6-4ee5-9d59-7aa43d325b71","Type":"ContainerDied","Data":"6ac9bb1fd81282b0a1c6bdb0d059a6a3560d3710c8a1d00a1a936867d8053db9"} Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.942478 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6bg" event={"ID":"6a34571c-5dc6-4ee5-9d59-7aa43d325b71","Type":"ContainerStarted","Data":"43f34f98ef1e015aa3ba221695da1f46205d00ce0a70af695b30367a555982b9"} Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.946547 4944 generic.go:334] "Generic (PLEG): container finished" podID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" containerID="bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629" exitCode=0 Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.946608 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s66h" event={"ID":"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf","Type":"ContainerDied","Data":"bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629"} Nov 24 08:54:53 crc kubenswrapper[4944]: I1124 08:54:53.950010 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcq2j" event={"ID":"eb7721e8-723b-435f-bb1c-e4b0fcf3372b","Type":"ContainerStarted","Data":"a45d285255c191fb0afde701452164214a94ce3c5758d26040644d9adeda753e"} Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.300471 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.464119 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38028fcd-3200-4382-967a-02637d30f33a-kubelet-dir\") pod \"38028fcd-3200-4382-967a-02637d30f33a\" (UID: \"38028fcd-3200-4382-967a-02637d30f33a\") " Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.464274 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38028fcd-3200-4382-967a-02637d30f33a-kube-api-access\") pod \"38028fcd-3200-4382-967a-02637d30f33a\" (UID: \"38028fcd-3200-4382-967a-02637d30f33a\") " Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.464405 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38028fcd-3200-4382-967a-02637d30f33a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "38028fcd-3200-4382-967a-02637d30f33a" (UID: "38028fcd-3200-4382-967a-02637d30f33a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.464842 4944 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38028fcd-3200-4382-967a-02637d30f33a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.472040 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38028fcd-3200-4382-967a-02637d30f33a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "38028fcd-3200-4382-967a-02637d30f33a" (UID: "38028fcd-3200-4382-967a-02637d30f33a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.567837 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38028fcd-3200-4382-967a-02637d30f33a-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.984532 4944 generic.go:334] "Generic (PLEG): container finished" podID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerID="f9bcc6fe81eb63d20fed2a6a6bf14d0274b608e2a490d2b5777283dbe3389819" exitCode=0 Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.984612 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcq2j" event={"ID":"eb7721e8-723b-435f-bb1c-e4b0fcf3372b","Type":"ContainerDied","Data":"f9bcc6fe81eb63d20fed2a6a6bf14d0274b608e2a490d2b5777283dbe3389819"} Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.991308 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"38028fcd-3200-4382-967a-02637d30f33a","Type":"ContainerDied","Data":"bc92b10816338e17bb60f27e416914feb3281596c66570af41f6451fa865090e"} Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.991358 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc92b10816338e17bb60f27e416914feb3281596c66570af41f6451fa865090e" Nov 24 08:54:54 crc kubenswrapper[4944]: I1124 08:54:54.991422 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.192692 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 08:54:55 crc kubenswrapper[4944]: E1124 08:54:55.193135 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38028fcd-3200-4382-967a-02637d30f33a" containerName="pruner" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.193157 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="38028fcd-3200-4382-967a-02637d30f33a" containerName="pruner" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.193314 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="38028fcd-3200-4382-967a-02637d30f33a" containerName="pruner" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.193779 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.193883 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.200209 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.221540 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.278857 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0125294b-18cf-4957-923a-d71d820a2472-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0125294b-18cf-4957-923a-d71d820a2472\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.278942 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0125294b-18cf-4957-923a-d71d820a2472-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0125294b-18cf-4957-923a-d71d820a2472\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.386822 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0125294b-18cf-4957-923a-d71d820a2472-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0125294b-18cf-4957-923a-d71d820a2472\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.386971 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0125294b-18cf-4957-923a-d71d820a2472-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0125294b-18cf-4957-923a-d71d820a2472\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.387000 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0125294b-18cf-4957-923a-d71d820a2472-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0125294b-18cf-4957-923a-d71d820a2472\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.418488 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0125294b-18cf-4957-923a-d71d820a2472-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0125294b-18cf-4957-923a-d71d820a2472\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.517869 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:54:55 crc kubenswrapper[4944]: I1124 08:54:55.938536 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 08:54:56 crc kubenswrapper[4944]: I1124 08:54:56.027653 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0125294b-18cf-4957-923a-d71d820a2472","Type":"ContainerStarted","Data":"a7f41d677e5866963a55d8d4e4869f43ce049c997fc1ad433582e59c6fd0d633"} Nov 24 08:54:56 crc kubenswrapper[4944]: I1124 08:54:56.601545 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-68hsd" Nov 24 08:54:57 crc kubenswrapper[4944]: I1124 08:54:57.070253 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0125294b-18cf-4957-923a-d71d820a2472","Type":"ContainerStarted","Data":"0187e01864fae5806b7c29c8ef7959092087019e097a27199688440310ec9393"} Nov 24 08:54:57 crc kubenswrapper[4944]: I1124 08:54:57.091498 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.091480228 podStartE2EDuration="2.091480228s" podCreationTimestamp="2025-11-24 08:54:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:54:57.089577042 +0000 UTC m=+157.624017514" watchObservedRunningTime="2025-11-24 08:54:57.091480228 +0000 UTC m=+157.625920690" Nov 24 08:54:57 crc kubenswrapper[4944]: I1124 08:54:57.516259 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:54:58 crc kubenswrapper[4944]: I1124 08:54:58.087189 4944 generic.go:334] "Generic (PLEG): container finished" podID="0125294b-18cf-4957-923a-d71d820a2472" containerID="0187e01864fae5806b7c29c8ef7959092087019e097a27199688440310ec9393" exitCode=0 Nov 24 08:54:58 crc kubenswrapper[4944]: I1124 08:54:58.087313 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0125294b-18cf-4957-923a-d71d820a2472","Type":"ContainerDied","Data":"0187e01864fae5806b7c29c8ef7959092087019e097a27199688440310ec9393"} Nov 24 08:54:59 crc kubenswrapper[4944]: I1124 08:54:59.250416 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 08:55:00 crc kubenswrapper[4944]: I1124 08:55:00.741317 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:55:00 crc kubenswrapper[4944]: I1124 08:55:00.745404 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 08:55:01 crc kubenswrapper[4944]: I1124 08:55:01.569857 4944 patch_prober.go:28] interesting pod/downloads-7954f5f757-9tglc container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:55:01 crc kubenswrapper[4944]: I1124 08:55:01.569915 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9tglc" podUID="bbaf6605-39cb-41a0-900a-b2a87a00751e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:55:01 crc kubenswrapper[4944]: I1124 08:55:01.570009 4944 patch_prober.go:28] interesting pod/downloads-7954f5f757-9tglc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:55:01 crc kubenswrapper[4944]: I1124 08:55:01.570078 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9tglc" podUID="bbaf6605-39cb-41a0-900a-b2a87a00751e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:55:01 crc kubenswrapper[4944]: I1124 08:55:01.638088 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:55:01 crc kubenswrapper[4944]: I1124 08:55:01.649792 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55990e0c-a6ff-4b18-8b8d-f27542761408-metrics-certs\") pod \"network-metrics-daemon-jmkb9\" (UID: \"55990e0c-a6ff-4b18-8b8d-f27542761408\") " pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:55:01 crc kubenswrapper[4944]: I1124 08:55:01.893186 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jmkb9" Nov 24 08:55:04 crc kubenswrapper[4944]: I1124 08:55:04.974703 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:55:05 crc kubenswrapper[4944]: I1124 08:55:05.130494 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0125294b-18cf-4957-923a-d71d820a2472-kube-api-access\") pod \"0125294b-18cf-4957-923a-d71d820a2472\" (UID: \"0125294b-18cf-4957-923a-d71d820a2472\") " Nov 24 08:55:05 crc kubenswrapper[4944]: I1124 08:55:05.130555 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0125294b-18cf-4957-923a-d71d820a2472-kubelet-dir\") pod \"0125294b-18cf-4957-923a-d71d820a2472\" (UID: \"0125294b-18cf-4957-923a-d71d820a2472\") " Nov 24 08:55:05 crc kubenswrapper[4944]: I1124 08:55:05.130752 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0125294b-18cf-4957-923a-d71d820a2472-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0125294b-18cf-4957-923a-d71d820a2472" (UID: "0125294b-18cf-4957-923a-d71d820a2472"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:55:05 crc kubenswrapper[4944]: I1124 08:55:05.135957 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0125294b-18cf-4957-923a-d71d820a2472-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0125294b-18cf-4957-923a-d71d820a2472" (UID: "0125294b-18cf-4957-923a-d71d820a2472"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:55:05 crc kubenswrapper[4944]: I1124 08:55:05.140376 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0125294b-18cf-4957-923a-d71d820a2472","Type":"ContainerDied","Data":"a7f41d677e5866963a55d8d4e4869f43ce049c997fc1ad433582e59c6fd0d633"} Nov 24 08:55:05 crc kubenswrapper[4944]: I1124 08:55:05.140416 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7f41d677e5866963a55d8d4e4869f43ce049c997fc1ad433582e59c6fd0d633" Nov 24 08:55:05 crc kubenswrapper[4944]: I1124 08:55:05.140422 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:55:05 crc kubenswrapper[4944]: I1124 08:55:05.231396 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0125294b-18cf-4957-923a-d71d820a2472-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:05 crc kubenswrapper[4944]: I1124 08:55:05.231438 4944 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0125294b-18cf-4957-923a-d71d820a2472-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:10 crc kubenswrapper[4944]: I1124 08:55:10.712492 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 08:55:11 crc kubenswrapper[4944]: I1124 08:55:11.589869 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-9tglc" Nov 24 08:55:17 crc kubenswrapper[4944]: I1124 08:55:17.236205 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jmkb9"] Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.218478 4944 generic.go:334] "Generic (PLEG): container finished" podID="32b94450-1b99-4fc9-aac8-658628a2195f" containerID="c09928912211f9e65062ca36ab486e44fdb1ca6b6834c4e1058927a7b0fbd4c1" exitCode=0 Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.218660 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6dlz8" event={"ID":"32b94450-1b99-4fc9-aac8-658628a2195f","Type":"ContainerDied","Data":"c09928912211f9e65062ca36ab486e44fdb1ca6b6834c4e1058927a7b0fbd4c1"} Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.221383 4944 generic.go:334] "Generic (PLEG): container finished" podID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" containerID="4d8e8c9ce00eae831eee8a47c0da32329b2a4d93660ee002dd0631608094d5fd" exitCode=0 Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.221444 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s66h" event={"ID":"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf","Type":"ContainerDied","Data":"4d8e8c9ce00eae831eee8a47c0da32329b2a4d93660ee002dd0631608094d5fd"} Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.223306 4944 generic.go:334] "Generic (PLEG): container finished" podID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerID="803a9195c2e6381720b44030b58d68ce33d685d559ad2cf141e3ca3af763aad4" exitCode=0 Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.223363 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcq2j" event={"ID":"eb7721e8-723b-435f-bb1c-e4b0fcf3372b","Type":"ContainerDied","Data":"803a9195c2e6381720b44030b58d68ce33d685d559ad2cf141e3ca3af763aad4"} Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.227714 4944 generic.go:334] "Generic (PLEG): container finished" podID="c8249efc-c3ac-4527-8394-e6e0aff9d457" containerID="7aea9faba41863011206d50fe384a02b9c8cddd84d5214a3d9e8b76670217a78" exitCode=0 Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.227781 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8cs79" event={"ID":"c8249efc-c3ac-4527-8394-e6e0aff9d457","Type":"ContainerDied","Data":"7aea9faba41863011206d50fe384a02b9c8cddd84d5214a3d9e8b76670217a78"} Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.229159 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" event={"ID":"55990e0c-a6ff-4b18-8b8d-f27542761408","Type":"ContainerStarted","Data":"237ee38746c5eb85e5769c17e1eb619a5481766f15429fd6f74086ee66aadd46"} Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.229203 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" event={"ID":"55990e0c-a6ff-4b18-8b8d-f27542761408","Type":"ContainerStarted","Data":"14bcd81f7efaa4454e3b7f8e51576944dd4e702abe4e421fd3641ad8af906fa5"} Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.230720 4944 generic.go:334] "Generic (PLEG): container finished" podID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerID="80af712ab259b6f36b3ba2450cd0ce1001e5a0be3251e2d614e5b7fff99be199" exitCode=0 Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.230766 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6bg" event={"ID":"6a34571c-5dc6-4ee5-9d59-7aa43d325b71","Type":"ContainerDied","Data":"80af712ab259b6f36b3ba2450cd0ce1001e5a0be3251e2d614e5b7fff99be199"} Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.255474 4944 generic.go:334] "Generic (PLEG): container finished" podID="9f532fb4-a915-4cf7-a64a-887eadfb1710" containerID="d0addb826471836eb3c9b89984301c9c815db0e7b72526cafdd9ac9c613e08f4" exitCode=0 Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.255491 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sphtd" event={"ID":"9f532fb4-a915-4cf7-a64a-887eadfb1710","Type":"ContainerDied","Data":"d0addb826471836eb3c9b89984301c9c815db0e7b72526cafdd9ac9c613e08f4"} Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.261381 4944 generic.go:334] "Generic (PLEG): container finished" podID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" containerID="2b5a5a6d9da6baa0dbc47975990bec6e85b21f7f458135fd2f21a206c06d4988" exitCode=0 Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.261439 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxw2" event={"ID":"a6587b50-bc2c-407f-b0f1-f22b6ea202cc","Type":"ContainerDied","Data":"2b5a5a6d9da6baa0dbc47975990bec6e85b21f7f458135fd2f21a206c06d4988"} Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.269940 4944 generic.go:334] "Generic (PLEG): container finished" podID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" containerID="65a1e742117334f183f9e9df5dc8c25ce827467e63db686028cb4580956a6759" exitCode=0 Nov 24 08:55:18 crc kubenswrapper[4944]: I1124 08:55:18.269994 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vsqrm" event={"ID":"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab","Type":"ContainerDied","Data":"65a1e742117334f183f9e9df5dc8c25ce827467e63db686028cb4580956a6759"} Nov 24 08:55:19 crc kubenswrapper[4944]: I1124 08:55:19.277265 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jmkb9" event={"ID":"55990e0c-a6ff-4b18-8b8d-f27542761408","Type":"ContainerStarted","Data":"3ffc4531ce60e814966d73e08e3fba74a69d073c8552698c691fc89e1ea7d989"} Nov 24 08:55:21 crc kubenswrapper[4944]: I1124 08:55:21.289106 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcq2j" event={"ID":"eb7721e8-723b-435f-bb1c-e4b0fcf3372b","Type":"ContainerStarted","Data":"e4243cd5bdb95ae200e7568e09b23f88d218fa7a62f9ad1197a8a21d0a6d79e8"} Nov 24 08:55:21 crc kubenswrapper[4944]: I1124 08:55:21.311332 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-jmkb9" podStartSLOduration=162.311315912 podStartE2EDuration="2m42.311315912s" podCreationTimestamp="2025-11-24 08:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:55:19.298445203 +0000 UTC m=+179.832885705" watchObservedRunningTime="2025-11-24 08:55:21.311315912 +0000 UTC m=+181.845756374" Nov 24 08:55:21 crc kubenswrapper[4944]: I1124 08:55:21.569897 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6j56z" Nov 24 08:55:21 crc kubenswrapper[4944]: I1124 08:55:21.584453 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pcq2j" podStartSLOduration=3.183817222 podStartE2EDuration="28.584431646s" podCreationTimestamp="2025-11-24 08:54:53 +0000 UTC" firstStartedPulling="2025-11-24 08:54:54.98877244 +0000 UTC m=+155.523212902" lastFinishedPulling="2025-11-24 08:55:20.389386864 +0000 UTC m=+180.923827326" observedRunningTime="2025-11-24 08:55:21.311686492 +0000 UTC m=+181.846126954" watchObservedRunningTime="2025-11-24 08:55:21.584431646 +0000 UTC m=+182.118872108" Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.296471 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vsqrm" event={"ID":"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab","Type":"ContainerStarted","Data":"7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed"} Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.298891 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8cs79" event={"ID":"c8249efc-c3ac-4527-8394-e6e0aff9d457","Type":"ContainerStarted","Data":"82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae"} Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.302383 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6bg" event={"ID":"6a34571c-5dc6-4ee5-9d59-7aa43d325b71","Type":"ContainerStarted","Data":"8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c"} Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.304928 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sphtd" event={"ID":"9f532fb4-a915-4cf7-a64a-887eadfb1710","Type":"ContainerStarted","Data":"481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076"} Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.306937 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6dlz8" event={"ID":"32b94450-1b99-4fc9-aac8-658628a2195f","Type":"ContainerStarted","Data":"eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e"} Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.309311 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxw2" event={"ID":"a6587b50-bc2c-407f-b0f1-f22b6ea202cc","Type":"ContainerStarted","Data":"d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164"} Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.312304 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s66h" event={"ID":"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf","Type":"ContainerStarted","Data":"433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582"} Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.342027 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vsqrm" podStartSLOduration=2.596680847 podStartE2EDuration="33.341992168s" podCreationTimestamp="2025-11-24 08:54:49 +0000 UTC" firstStartedPulling="2025-11-24 08:54:50.798615016 +0000 UTC m=+151.333055478" lastFinishedPulling="2025-11-24 08:55:21.543926337 +0000 UTC m=+182.078366799" observedRunningTime="2025-11-24 08:55:22.317393561 +0000 UTC m=+182.851834023" watchObservedRunningTime="2025-11-24 08:55:22.341992168 +0000 UTC m=+182.876432630" Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.342648 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9q6bg" podStartSLOduration=2.796976762 podStartE2EDuration="30.342638748s" podCreationTimestamp="2025-11-24 08:54:52 +0000 UTC" firstStartedPulling="2025-11-24 08:54:53.944027766 +0000 UTC m=+154.478468228" lastFinishedPulling="2025-11-24 08:55:21.489689752 +0000 UTC m=+182.024130214" observedRunningTime="2025-11-24 08:55:22.340265978 +0000 UTC m=+182.874706450" watchObservedRunningTime="2025-11-24 08:55:22.342638748 +0000 UTC m=+182.877079220" Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.374985 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6dlz8" podStartSLOduration=2.524721935 podStartE2EDuration="31.374969475s" podCreationTimestamp="2025-11-24 08:54:51 +0000 UTC" firstStartedPulling="2025-11-24 08:54:52.916688038 +0000 UTC m=+153.451128500" lastFinishedPulling="2025-11-24 08:55:21.766935578 +0000 UTC m=+182.301376040" observedRunningTime="2025-11-24 08:55:22.368453052 +0000 UTC m=+182.902893514" watchObservedRunningTime="2025-11-24 08:55:22.374969475 +0000 UTC m=+182.909409937" Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.438315 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sphtd" podStartSLOduration=2.5840785840000002 podStartE2EDuration="33.438298899s" podCreationTimestamp="2025-11-24 08:54:49 +0000 UTC" firstStartedPulling="2025-11-24 08:54:50.780157289 +0000 UTC m=+151.314597751" lastFinishedPulling="2025-11-24 08:55:21.634377604 +0000 UTC m=+182.168818066" observedRunningTime="2025-11-24 08:55:22.435175287 +0000 UTC m=+182.969615749" watchObservedRunningTime="2025-11-24 08:55:22.438298899 +0000 UTC m=+182.972739361" Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.479360 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9cxw2" podStartSLOduration=2.460452093 podStartE2EDuration="32.479339214s" podCreationTimestamp="2025-11-24 08:54:50 +0000 UTC" firstStartedPulling="2025-11-24 08:54:51.829993733 +0000 UTC m=+152.364434195" lastFinishedPulling="2025-11-24 08:55:21.848880864 +0000 UTC m=+182.383321316" observedRunningTime="2025-11-24 08:55:22.479277052 +0000 UTC m=+183.013717514" watchObservedRunningTime="2025-11-24 08:55:22.479339214 +0000 UTC m=+183.013779676" Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.502220 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4s66h" podStartSLOduration=2.444951354 podStartE2EDuration="31.502201401s" podCreationTimestamp="2025-11-24 08:54:51 +0000 UTC" firstStartedPulling="2025-11-24 08:54:52.916712919 +0000 UTC m=+153.451153381" lastFinishedPulling="2025-11-24 08:55:21.973962966 +0000 UTC m=+182.508403428" observedRunningTime="2025-11-24 08:55:22.501770038 +0000 UTC m=+183.036210500" watchObservedRunningTime="2025-11-24 08:55:22.502201401 +0000 UTC m=+183.036641873" Nov 24 08:55:22 crc kubenswrapper[4944]: I1124 08:55:22.536758 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8cs79" podStartSLOduration=2.800256862 podStartE2EDuration="33.536742543s" podCreationTimestamp="2025-11-24 08:54:49 +0000 UTC" firstStartedPulling="2025-11-24 08:54:50.807278692 +0000 UTC m=+151.341719154" lastFinishedPulling="2025-11-24 08:55:21.543764373 +0000 UTC m=+182.078204835" observedRunningTime="2025-11-24 08:55:22.535436164 +0000 UTC m=+183.069876626" watchObservedRunningTime="2025-11-24 08:55:22.536742543 +0000 UTC m=+183.071183005" Nov 24 08:55:23 crc kubenswrapper[4944]: I1124 08:55:23.079616 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:55:23 crc kubenswrapper[4944]: I1124 08:55:23.080001 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:55:23 crc kubenswrapper[4944]: I1124 08:55:23.392808 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:55:23 crc kubenswrapper[4944]: I1124 08:55:23.392910 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:55:23 crc kubenswrapper[4944]: I1124 08:55:23.549206 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:55:23 crc kubenswrapper[4944]: I1124 08:55:23.549603 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:55:24 crc kubenswrapper[4944]: I1124 08:55:24.435289 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9q6bg" podUID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerName="registry-server" probeResult="failure" output=< Nov 24 08:55:24 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 08:55:24 crc kubenswrapper[4944]: > Nov 24 08:55:24 crc kubenswrapper[4944]: I1124 08:55:24.440639 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pcq2j" podUID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerName="registry-server" probeResult="failure" output=< Nov 24 08:55:24 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 08:55:24 crc kubenswrapper[4944]: > Nov 24 08:55:27 crc kubenswrapper[4944]: I1124 08:55:27.555584 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:55:29 crc kubenswrapper[4944]: I1124 08:55:29.754466 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:55:29 crc kubenswrapper[4944]: I1124 08:55:29.755341 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:55:29 crc kubenswrapper[4944]: I1124 08:55:29.842947 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:55:29 crc kubenswrapper[4944]: I1124 08:55:29.954617 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:55:29 crc kubenswrapper[4944]: I1124 08:55:29.955145 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:55:30 crc kubenswrapper[4944]: I1124 08:55:30.002604 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:55:30 crc kubenswrapper[4944]: I1124 08:55:30.158795 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:55:30 crc kubenswrapper[4944]: I1124 08:55:30.159748 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:55:30 crc kubenswrapper[4944]: I1124 08:55:30.234704 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-h6vpv"] Nov 24 08:55:30 crc kubenswrapper[4944]: I1124 08:55:30.238991 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:55:30 crc kubenswrapper[4944]: I1124 08:55:30.404453 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:55:30 crc kubenswrapper[4944]: I1124 08:55:30.423853 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:55:30 crc kubenswrapper[4944]: I1124 08:55:30.426609 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:55:30 crc kubenswrapper[4944]: I1124 08:55:30.426659 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:55:30 crc kubenswrapper[4944]: I1124 08:55:30.429176 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:55:30 crc kubenswrapper[4944]: I1124 08:55:30.503293 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:55:31 crc kubenswrapper[4944]: I1124 08:55:31.398740 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:55:31 crc kubenswrapper[4944]: I1124 08:55:31.762966 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:55:31 crc kubenswrapper[4944]: I1124 08:55:31.764131 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:55:31 crc kubenswrapper[4944]: I1124 08:55:31.797791 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:55:32 crc kubenswrapper[4944]: I1124 08:55:32.157031 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:55:32 crc kubenswrapper[4944]: I1124 08:55:32.158838 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:55:32 crc kubenswrapper[4944]: I1124 08:55:32.198021 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:55:32 crc kubenswrapper[4944]: I1124 08:55:32.406253 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:55:32 crc kubenswrapper[4944]: I1124 08:55:32.412254 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:55:32 crc kubenswrapper[4944]: I1124 08:55:32.458753 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vsqrm"] Nov 24 08:55:32 crc kubenswrapper[4944]: I1124 08:55:32.659166 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9cxw2"] Nov 24 08:55:33 crc kubenswrapper[4944]: I1124 08:55:33.121818 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:55:33 crc kubenswrapper[4944]: I1124 08:55:33.160274 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:55:33 crc kubenswrapper[4944]: I1124 08:55:33.368164 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9cxw2" podUID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" containerName="registry-server" containerID="cri-o://d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164" gracePeriod=2 Nov 24 08:55:33 crc kubenswrapper[4944]: I1124 08:55:33.368677 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vsqrm" podUID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" containerName="registry-server" containerID="cri-o://7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed" gracePeriod=2 Nov 24 08:55:33 crc kubenswrapper[4944]: I1124 08:55:33.431803 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:55:33 crc kubenswrapper[4944]: I1124 08:55:33.480474 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:55:33 crc kubenswrapper[4944]: I1124 08:55:33.836904 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:55:33 crc kubenswrapper[4944]: I1124 08:55:33.892065 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.015749 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-catalog-content\") pod \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.015813 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-catalog-content\") pod \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.015872 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-utilities\") pod \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.015918 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-utilities\") pod \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.016005 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxhmm\" (UniqueName: \"kubernetes.io/projected/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-kube-api-access-jxhmm\") pod \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\" (UID: \"a6587b50-bc2c-407f-b0f1-f22b6ea202cc\") " Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.016027 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndb9f\" (UniqueName: \"kubernetes.io/projected/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-kube-api-access-ndb9f\") pod \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\" (UID: \"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab\") " Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.016577 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-utilities" (OuterVolumeSpecName: "utilities") pod "a6587b50-bc2c-407f-b0f1-f22b6ea202cc" (UID: "a6587b50-bc2c-407f-b0f1-f22b6ea202cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.017436 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-utilities" (OuterVolumeSpecName: "utilities") pod "a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" (UID: "a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.022004 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-kube-api-access-ndb9f" (OuterVolumeSpecName: "kube-api-access-ndb9f") pod "a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" (UID: "a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab"). InnerVolumeSpecName "kube-api-access-ndb9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.022497 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-kube-api-access-jxhmm" (OuterVolumeSpecName: "kube-api-access-jxhmm") pod "a6587b50-bc2c-407f-b0f1-f22b6ea202cc" (UID: "a6587b50-bc2c-407f-b0f1-f22b6ea202cc"). InnerVolumeSpecName "kube-api-access-jxhmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.071536 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6587b50-bc2c-407f-b0f1-f22b6ea202cc" (UID: "a6587b50-bc2c-407f-b0f1-f22b6ea202cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.071791 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" (UID: "a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.117740 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxhmm\" (UniqueName: \"kubernetes.io/projected/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-kube-api-access-jxhmm\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.117791 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndb9f\" (UniqueName: \"kubernetes.io/projected/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-kube-api-access-ndb9f\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.117801 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.117810 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.117821 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6587b50-bc2c-407f-b0f1-f22b6ea202cc-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.117831 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.376549 4944 generic.go:334] "Generic (PLEG): container finished" podID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" containerID="d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164" exitCode=0 Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.376669 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cxw2" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.376662 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxw2" event={"ID":"a6587b50-bc2c-407f-b0f1-f22b6ea202cc","Type":"ContainerDied","Data":"d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164"} Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.376856 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxw2" event={"ID":"a6587b50-bc2c-407f-b0f1-f22b6ea202cc","Type":"ContainerDied","Data":"c396ee2079dd470d2cd231c3dc183261c6c36467b0e807b9783c27770276463c"} Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.376894 4944 scope.go:117] "RemoveContainer" containerID="d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.385256 4944 generic.go:334] "Generic (PLEG): container finished" podID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" containerID="7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed" exitCode=0 Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.385352 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vsqrm" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.385387 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vsqrm" event={"ID":"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab","Type":"ContainerDied","Data":"7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed"} Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.385603 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vsqrm" event={"ID":"a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab","Type":"ContainerDied","Data":"81bb06da389a25a2402232d390c51cdec29e7cdf5a908349aa264cad993bd339"} Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.396878 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9cxw2"] Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.399523 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9cxw2"] Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.401559 4944 scope.go:117] "RemoveContainer" containerID="2b5a5a6d9da6baa0dbc47975990bec6e85b21f7f458135fd2f21a206c06d4988" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.409939 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vsqrm"] Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.415880 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vsqrm"] Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.433308 4944 scope.go:117] "RemoveContainer" containerID="e9175e625095da4269400c2f8310ef0e9dce4efceb2006987379432ef18e3473" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.459807 4944 scope.go:117] "RemoveContainer" containerID="d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164" Nov 24 08:55:34 crc kubenswrapper[4944]: E1124 08:55:34.460346 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164\": container with ID starting with d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164 not found: ID does not exist" containerID="d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.460395 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164"} err="failed to get container status \"d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164\": rpc error: code = NotFound desc = could not find container \"d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164\": container with ID starting with d4e152f75418b9f5f35d74810f2ecb6620bcf02363b57c63aac74db37866f164 not found: ID does not exist" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.460441 4944 scope.go:117] "RemoveContainer" containerID="2b5a5a6d9da6baa0dbc47975990bec6e85b21f7f458135fd2f21a206c06d4988" Nov 24 08:55:34 crc kubenswrapper[4944]: E1124 08:55:34.460907 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b5a5a6d9da6baa0dbc47975990bec6e85b21f7f458135fd2f21a206c06d4988\": container with ID starting with 2b5a5a6d9da6baa0dbc47975990bec6e85b21f7f458135fd2f21a206c06d4988 not found: ID does not exist" containerID="2b5a5a6d9da6baa0dbc47975990bec6e85b21f7f458135fd2f21a206c06d4988" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.460983 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b5a5a6d9da6baa0dbc47975990bec6e85b21f7f458135fd2f21a206c06d4988"} err="failed to get container status \"2b5a5a6d9da6baa0dbc47975990bec6e85b21f7f458135fd2f21a206c06d4988\": rpc error: code = NotFound desc = could not find container \"2b5a5a6d9da6baa0dbc47975990bec6e85b21f7f458135fd2f21a206c06d4988\": container with ID starting with 2b5a5a6d9da6baa0dbc47975990bec6e85b21f7f458135fd2f21a206c06d4988 not found: ID does not exist" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.461022 4944 scope.go:117] "RemoveContainer" containerID="e9175e625095da4269400c2f8310ef0e9dce4efceb2006987379432ef18e3473" Nov 24 08:55:34 crc kubenswrapper[4944]: E1124 08:55:34.461651 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9175e625095da4269400c2f8310ef0e9dce4efceb2006987379432ef18e3473\": container with ID starting with e9175e625095da4269400c2f8310ef0e9dce4efceb2006987379432ef18e3473 not found: ID does not exist" containerID="e9175e625095da4269400c2f8310ef0e9dce4efceb2006987379432ef18e3473" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.461682 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9175e625095da4269400c2f8310ef0e9dce4efceb2006987379432ef18e3473"} err="failed to get container status \"e9175e625095da4269400c2f8310ef0e9dce4efceb2006987379432ef18e3473\": rpc error: code = NotFound desc = could not find container \"e9175e625095da4269400c2f8310ef0e9dce4efceb2006987379432ef18e3473\": container with ID starting with e9175e625095da4269400c2f8310ef0e9dce4efceb2006987379432ef18e3473 not found: ID does not exist" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.461707 4944 scope.go:117] "RemoveContainer" containerID="7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.476508 4944 scope.go:117] "RemoveContainer" containerID="65a1e742117334f183f9e9df5dc8c25ce827467e63db686028cb4580956a6759" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.490779 4944 scope.go:117] "RemoveContainer" containerID="2f5c19aa2d44475dbec454564398e9bceeceb6439b4f31da4a5b53386cc28ec4" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.504301 4944 scope.go:117] "RemoveContainer" containerID="7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed" Nov 24 08:55:34 crc kubenswrapper[4944]: E1124 08:55:34.504984 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed\": container with ID starting with 7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed not found: ID does not exist" containerID="7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.505020 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed"} err="failed to get container status \"7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed\": rpc error: code = NotFound desc = could not find container \"7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed\": container with ID starting with 7726fd7e289d2798394d930ae68d74edc0b7562296740c77badf13820ddeafed not found: ID does not exist" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.505059 4944 scope.go:117] "RemoveContainer" containerID="65a1e742117334f183f9e9df5dc8c25ce827467e63db686028cb4580956a6759" Nov 24 08:55:34 crc kubenswrapper[4944]: E1124 08:55:34.506148 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65a1e742117334f183f9e9df5dc8c25ce827467e63db686028cb4580956a6759\": container with ID starting with 65a1e742117334f183f9e9df5dc8c25ce827467e63db686028cb4580956a6759 not found: ID does not exist" containerID="65a1e742117334f183f9e9df5dc8c25ce827467e63db686028cb4580956a6759" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.506201 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65a1e742117334f183f9e9df5dc8c25ce827467e63db686028cb4580956a6759"} err="failed to get container status \"65a1e742117334f183f9e9df5dc8c25ce827467e63db686028cb4580956a6759\": rpc error: code = NotFound desc = could not find container \"65a1e742117334f183f9e9df5dc8c25ce827467e63db686028cb4580956a6759\": container with ID starting with 65a1e742117334f183f9e9df5dc8c25ce827467e63db686028cb4580956a6759 not found: ID does not exist" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.506233 4944 scope.go:117] "RemoveContainer" containerID="2f5c19aa2d44475dbec454564398e9bceeceb6439b4f31da4a5b53386cc28ec4" Nov 24 08:55:34 crc kubenswrapper[4944]: E1124 08:55:34.507460 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f5c19aa2d44475dbec454564398e9bceeceb6439b4f31da4a5b53386cc28ec4\": container with ID starting with 2f5c19aa2d44475dbec454564398e9bceeceb6439b4f31da4a5b53386cc28ec4 not found: ID does not exist" containerID="2f5c19aa2d44475dbec454564398e9bceeceb6439b4f31da4a5b53386cc28ec4" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.507496 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f5c19aa2d44475dbec454564398e9bceeceb6439b4f31da4a5b53386cc28ec4"} err="failed to get container status \"2f5c19aa2d44475dbec454564398e9bceeceb6439b4f31da4a5b53386cc28ec4\": rpc error: code = NotFound desc = could not find container \"2f5c19aa2d44475dbec454564398e9bceeceb6439b4f31da4a5b53386cc28ec4\": container with ID starting with 2f5c19aa2d44475dbec454564398e9bceeceb6439b4f31da4a5b53386cc28ec4 not found: ID does not exist" Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.858662 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s66h"] Nov 24 08:55:34 crc kubenswrapper[4944]: I1124 08:55:34.858890 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4s66h" podUID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" containerName="registry-server" containerID="cri-o://433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582" gracePeriod=2 Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.231123 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.332651 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-utilities\") pod \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.332781 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-catalog-content\") pod \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.332826 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b68fl\" (UniqueName: \"kubernetes.io/projected/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-kube-api-access-b68fl\") pod \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\" (UID: \"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf\") " Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.334204 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-utilities" (OuterVolumeSpecName: "utilities") pod "bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" (UID: "bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.336541 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-kube-api-access-b68fl" (OuterVolumeSpecName: "kube-api-access-b68fl") pod "bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" (UID: "bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf"). InnerVolumeSpecName "kube-api-access-b68fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.353090 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" (UID: "bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.396252 4944 generic.go:334] "Generic (PLEG): container finished" podID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" containerID="433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582" exitCode=0 Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.396334 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s66h" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.396359 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s66h" event={"ID":"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf","Type":"ContainerDied","Data":"433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582"} Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.396431 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s66h" event={"ID":"bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf","Type":"ContainerDied","Data":"1bfc78bbd7b596e3664d97e815b0f853e8be6f7aec57808ae3b524fc8129f52f"} Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.396453 4944 scope.go:117] "RemoveContainer" containerID="433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.412962 4944 scope.go:117] "RemoveContainer" containerID="4d8e8c9ce00eae831eee8a47c0da32329b2a4d93660ee002dd0631608094d5fd" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.422657 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s66h"] Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.428963 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s66h"] Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.434653 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.434675 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b68fl\" (UniqueName: \"kubernetes.io/projected/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-kube-api-access-b68fl\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.434686 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.449039 4944 scope.go:117] "RemoveContainer" containerID="bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.466730 4944 scope.go:117] "RemoveContainer" containerID="433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582" Nov 24 08:55:35 crc kubenswrapper[4944]: E1124 08:55:35.467203 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582\": container with ID starting with 433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582 not found: ID does not exist" containerID="433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.467241 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582"} err="failed to get container status \"433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582\": rpc error: code = NotFound desc = could not find container \"433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582\": container with ID starting with 433b3c0dacef343daca1967d8b5321a4f86c96b07d51e98014c821b0f98bd582 not found: ID does not exist" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.467268 4944 scope.go:117] "RemoveContainer" containerID="4d8e8c9ce00eae831eee8a47c0da32329b2a4d93660ee002dd0631608094d5fd" Nov 24 08:55:35 crc kubenswrapper[4944]: E1124 08:55:35.467617 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d8e8c9ce00eae831eee8a47c0da32329b2a4d93660ee002dd0631608094d5fd\": container with ID starting with 4d8e8c9ce00eae831eee8a47c0da32329b2a4d93660ee002dd0631608094d5fd not found: ID does not exist" containerID="4d8e8c9ce00eae831eee8a47c0da32329b2a4d93660ee002dd0631608094d5fd" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.467644 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d8e8c9ce00eae831eee8a47c0da32329b2a4d93660ee002dd0631608094d5fd"} err="failed to get container status \"4d8e8c9ce00eae831eee8a47c0da32329b2a4d93660ee002dd0631608094d5fd\": rpc error: code = NotFound desc = could not find container \"4d8e8c9ce00eae831eee8a47c0da32329b2a4d93660ee002dd0631608094d5fd\": container with ID starting with 4d8e8c9ce00eae831eee8a47c0da32329b2a4d93660ee002dd0631608094d5fd not found: ID does not exist" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.467664 4944 scope.go:117] "RemoveContainer" containerID="bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629" Nov 24 08:55:35 crc kubenswrapper[4944]: E1124 08:55:35.468061 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629\": container with ID starting with bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629 not found: ID does not exist" containerID="bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629" Nov 24 08:55:35 crc kubenswrapper[4944]: I1124 08:55:35.468183 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629"} err="failed to get container status \"bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629\": rpc error: code = NotFound desc = could not find container \"bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629\": container with ID starting with bd44f23528a1b4916b1aaf5b44cbea6a30450eea5b72df24a89424f045a84629 not found: ID does not exist" Nov 24 08:55:36 crc kubenswrapper[4944]: I1124 08:55:36.282371 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" path="/var/lib/kubelet/pods/a6587b50-bc2c-407f-b0f1-f22b6ea202cc/volumes" Nov 24 08:55:36 crc kubenswrapper[4944]: I1124 08:55:36.282938 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" path="/var/lib/kubelet/pods/a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab/volumes" Nov 24 08:55:36 crc kubenswrapper[4944]: I1124 08:55:36.283547 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" path="/var/lib/kubelet/pods/bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf/volumes" Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.259457 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pcq2j"] Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.260841 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pcq2j" podUID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerName="registry-server" containerID="cri-o://e4243cd5bdb95ae200e7568e09b23f88d218fa7a62f9ad1197a8a21d0a6d79e8" gracePeriod=2 Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.411076 4944 generic.go:334] "Generic (PLEG): container finished" podID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerID="e4243cd5bdb95ae200e7568e09b23f88d218fa7a62f9ad1197a8a21d0a6d79e8" exitCode=0 Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.411088 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcq2j" event={"ID":"eb7721e8-723b-435f-bb1c-e4b0fcf3372b","Type":"ContainerDied","Data":"e4243cd5bdb95ae200e7568e09b23f88d218fa7a62f9ad1197a8a21d0a6d79e8"} Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.615934 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.666180 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-catalog-content\") pod \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.666275 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xntk6\" (UniqueName: \"kubernetes.io/projected/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-kube-api-access-xntk6\") pod \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.666337 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-utilities\") pod \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\" (UID: \"eb7721e8-723b-435f-bb1c-e4b0fcf3372b\") " Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.667930 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-utilities" (OuterVolumeSpecName: "utilities") pod "eb7721e8-723b-435f-bb1c-e4b0fcf3372b" (UID: "eb7721e8-723b-435f-bb1c-e4b0fcf3372b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.672207 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-kube-api-access-xntk6" (OuterVolumeSpecName: "kube-api-access-xntk6") pod "eb7721e8-723b-435f-bb1c-e4b0fcf3372b" (UID: "eb7721e8-723b-435f-bb1c-e4b0fcf3372b"). InnerVolumeSpecName "kube-api-access-xntk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.768266 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xntk6\" (UniqueName: \"kubernetes.io/projected/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-kube-api-access-xntk6\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.768308 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.780423 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb7721e8-723b-435f-bb1c-e4b0fcf3372b" (UID: "eb7721e8-723b-435f-bb1c-e4b0fcf3372b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:55:37 crc kubenswrapper[4944]: I1124 08:55:37.869152 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7721e8-723b-435f-bb1c-e4b0fcf3372b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:38 crc kubenswrapper[4944]: I1124 08:55:38.427846 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcq2j" event={"ID":"eb7721e8-723b-435f-bb1c-e4b0fcf3372b","Type":"ContainerDied","Data":"a45d285255c191fb0afde701452164214a94ce3c5758d26040644d9adeda753e"} Nov 24 08:55:38 crc kubenswrapper[4944]: I1124 08:55:38.427941 4944 scope.go:117] "RemoveContainer" containerID="e4243cd5bdb95ae200e7568e09b23f88d218fa7a62f9ad1197a8a21d0a6d79e8" Nov 24 08:55:38 crc kubenswrapper[4944]: I1124 08:55:38.427966 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcq2j" Nov 24 08:55:38 crc kubenswrapper[4944]: I1124 08:55:38.452020 4944 scope.go:117] "RemoveContainer" containerID="803a9195c2e6381720b44030b58d68ce33d685d559ad2cf141e3ca3af763aad4" Nov 24 08:55:38 crc kubenswrapper[4944]: I1124 08:55:38.454320 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pcq2j"] Nov 24 08:55:38 crc kubenswrapper[4944]: I1124 08:55:38.458352 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pcq2j"] Nov 24 08:55:38 crc kubenswrapper[4944]: I1124 08:55:38.468553 4944 scope.go:117] "RemoveContainer" containerID="f9bcc6fe81eb63d20fed2a6a6bf14d0274b608e2a490d2b5777283dbe3389819" Nov 24 08:55:40 crc kubenswrapper[4944]: I1124 08:55:40.284858 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" path="/var/lib/kubelet/pods/eb7721e8-723b-435f-bb1c-e4b0fcf3372b/volumes" Nov 24 08:55:53 crc kubenswrapper[4944]: I1124 08:55:53.548467 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:55:53 crc kubenswrapper[4944]: I1124 08:55:53.548986 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:55:53 crc kubenswrapper[4944]: I1124 08:55:53.549033 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:55:53 crc kubenswrapper[4944]: I1124 08:55:53.549647 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:55:53 crc kubenswrapper[4944]: I1124 08:55:53.549703 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83" gracePeriod=600 Nov 24 08:55:54 crc kubenswrapper[4944]: I1124 08:55:54.511002 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83" exitCode=0 Nov 24 08:55:54 crc kubenswrapper[4944]: I1124 08:55:54.511088 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83"} Nov 24 08:55:54 crc kubenswrapper[4944]: I1124 08:55:54.511297 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"152ff642f70a7f3a79f32127c3fc1fd67cb513fd122cd4fb960ccda5b2c27d78"} Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.271551 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" podUID="aeedd47e-ece2-4717-839b-755f1e1cd8e0" containerName="oauth-openshift" containerID="cri-o://1d7bcf30759d7b3a498a3f37641862f62b03336605dfd8af6b8a78981a692c83" gracePeriod=15 Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.517975 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeedd47e-ece2-4717-839b-755f1e1cd8e0" containerID="1d7bcf30759d7b3a498a3f37641862f62b03336605dfd8af6b8a78981a692c83" exitCode=0 Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.518089 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" event={"ID":"aeedd47e-ece2-4717-839b-755f1e1cd8e0","Type":"ContainerDied","Data":"1d7bcf30759d7b3a498a3f37641862f62b03336605dfd8af6b8a78981a692c83"} Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.581941 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.609748 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27"] Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610014 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610029 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610059 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" containerName="extract-utilities" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610066 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" containerName="extract-utilities" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610076 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0125294b-18cf-4957-923a-d71d820a2472" containerName="pruner" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610084 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0125294b-18cf-4957-923a-d71d820a2472" containerName="pruner" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610095 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" containerName="extract-content" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610102 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" containerName="extract-content" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610111 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" containerName="extract-content" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610117 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" containerName="extract-content" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610130 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610137 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610148 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerName="extract-utilities" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610155 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerName="extract-utilities" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610164 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerName="extract-content" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610171 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerName="extract-content" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610183 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" containerName="extract-utilities" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610190 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" containerName="extract-utilities" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610204 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610214 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610228 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeedd47e-ece2-4717-839b-755f1e1cd8e0" containerName="oauth-openshift" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610237 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeedd47e-ece2-4717-839b-755f1e1cd8e0" containerName="oauth-openshift" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610246 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" containerName="extract-content" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610252 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" containerName="extract-content" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610263 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610270 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: E1124 08:55:55.610281 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" containerName="extract-utilities" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610288 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" containerName="extract-utilities" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610390 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb7721e8-723b-435f-bb1c-e4b0fcf3372b" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610407 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a86d01c0-a24f-4b9c-a2ec-5fb1cd86d0ab" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610419 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeedd47e-ece2-4717-839b-755f1e1cd8e0" containerName="oauth-openshift" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610430 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc7b9c46-1cca-4ad4-8570-99ea7a5c52bf" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610439 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0125294b-18cf-4957-923a-d71d820a2472" containerName="pruner" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610448 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6587b50-bc2c-407f-b0f1-f22b6ea202cc" containerName="registry-server" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.610866 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.633556 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27"] Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.718584 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-router-certs\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.718664 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ms8j\" (UniqueName: \"kubernetes.io/projected/aeedd47e-ece2-4717-839b-755f1e1cd8e0-kube-api-access-4ms8j\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.718707 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-idp-0-file-data\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.718736 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-dir\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.718768 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-login\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.718809 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-cliconfig\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.718860 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-error\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.718886 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.718906 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-service-ca\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719000 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-serving-cert\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719032 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-trusted-ca-bundle\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719129 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-provider-selection\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719169 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-session\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719257 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-policies\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719321 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-ocp-branding-template\") pod \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\" (UID: \"aeedd47e-ece2-4717-839b-755f1e1cd8e0\") " Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719551 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-audit-policies\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719593 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719622 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719644 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmnw9\" (UniqueName: \"kubernetes.io/projected/f771e095-4b6d-4056-b498-a8472ad6d386-kube-api-access-vmnw9\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719670 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719715 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f771e095-4b6d-4056-b498-a8472ad6d386-audit-dir\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719749 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719773 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719800 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-template-login\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719823 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719857 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719902 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719929 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-template-error\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.719956 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-session\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.720007 4944 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.720534 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.720528 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.720586 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.721628 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.730775 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.731068 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeedd47e-ece2-4717-839b-755f1e1cd8e0-kube-api-access-4ms8j" (OuterVolumeSpecName: "kube-api-access-4ms8j") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "kube-api-access-4ms8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.731313 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.731901 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.732144 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.732413 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.732649 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.732881 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.736532 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "aeedd47e-ece2-4717-839b-755f1e1cd8e0" (UID: "aeedd47e-ece2-4717-839b-755f1e1cd8e0"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821133 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821201 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821231 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmnw9\" (UniqueName: \"kubernetes.io/projected/f771e095-4b6d-4056-b498-a8472ad6d386-kube-api-access-vmnw9\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821252 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821291 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f771e095-4b6d-4056-b498-a8472ad6d386-audit-dir\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821326 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821351 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821380 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-template-login\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821408 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821445 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821480 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821504 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-template-error\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821533 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-session\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821566 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-audit-policies\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821618 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821632 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821645 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ms8j\" (UniqueName: \"kubernetes.io/projected/aeedd47e-ece2-4717-839b-755f1e1cd8e0-kube-api-access-4ms8j\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821658 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821674 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821688 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821747 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821761 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821772 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821785 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821798 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821812 4944 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aeedd47e-ece2-4717-839b-755f1e1cd8e0-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.821824 4944 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aeedd47e-ece2-4717-839b-755f1e1cd8e0-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.822454 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.822801 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-audit-policies\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.822897 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f771e095-4b6d-4056-b498-a8472ad6d386-audit-dir\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.822945 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.823972 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.826498 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-template-error\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.826550 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.827809 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.827808 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-session\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.829616 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-template-login\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.829974 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.830302 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.830459 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f771e095-4b6d-4056-b498-a8472ad6d386-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.845928 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmnw9\" (UniqueName: \"kubernetes.io/projected/f771e095-4b6d-4056-b498-a8472ad6d386-kube-api-access-vmnw9\") pod \"oauth-openshift-5d7d85f8c7-4zj27\" (UID: \"f771e095-4b6d-4056-b498-a8472ad6d386\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:55 crc kubenswrapper[4944]: I1124 08:55:55.938881 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:56 crc kubenswrapper[4944]: I1124 08:55:56.130195 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27"] Nov 24 08:55:56 crc kubenswrapper[4944]: I1124 08:55:56.526263 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" event={"ID":"aeedd47e-ece2-4717-839b-755f1e1cd8e0","Type":"ContainerDied","Data":"aad46a88bb352e8961b434bef4f5404cae3a0d06e77631ac5c7ab3c1eff7a003"} Nov 24 08:55:56 crc kubenswrapper[4944]: I1124 08:55:56.527515 4944 scope.go:117] "RemoveContainer" containerID="1d7bcf30759d7b3a498a3f37641862f62b03336605dfd8af6b8a78981a692c83" Nov 24 08:55:56 crc kubenswrapper[4944]: I1124 08:55:56.526539 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-h6vpv" Nov 24 08:55:56 crc kubenswrapper[4944]: I1124 08:55:56.528021 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" event={"ID":"f771e095-4b6d-4056-b498-a8472ad6d386","Type":"ContainerStarted","Data":"e59a2fc39f3b3989bb9bd6cf5485fa08a6c67313652c06321a4734de4b8ad222"} Nov 24 08:55:56 crc kubenswrapper[4944]: I1124 08:55:56.528089 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" event={"ID":"f771e095-4b6d-4056-b498-a8472ad6d386","Type":"ContainerStarted","Data":"73465ec1cae3068d906b0a34b70a801fd990c69b412aee0d0d33a71a66402ed0"} Nov 24 08:55:56 crc kubenswrapper[4944]: I1124 08:55:56.528587 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:56 crc kubenswrapper[4944]: I1124 08:55:56.550655 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" podStartSLOduration=26.550631652 podStartE2EDuration="26.550631652s" podCreationTimestamp="2025-11-24 08:55:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:55:56.550326922 +0000 UTC m=+217.084767414" watchObservedRunningTime="2025-11-24 08:55:56.550631652 +0000 UTC m=+217.085072124" Nov 24 08:55:56 crc kubenswrapper[4944]: I1124 08:55:56.565431 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-h6vpv"] Nov 24 08:55:56 crc kubenswrapper[4944]: I1124 08:55:56.568763 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-h6vpv"] Nov 24 08:55:56 crc kubenswrapper[4944]: I1124 08:55:56.740852 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-4zj27" Nov 24 08:55:58 crc kubenswrapper[4944]: I1124 08:55:58.286915 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeedd47e-ece2-4717-839b-755f1e1cd8e0" path="/var/lib/kubelet/pods/aeedd47e-ece2-4717-839b-755f1e1cd8e0/volumes" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.011574 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sphtd"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.012435 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sphtd" podUID="9f532fb4-a915-4cf7-a64a-887eadfb1710" containerName="registry-server" containerID="cri-o://481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076" gracePeriod=30 Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.024886 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8cs79"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.025385 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8cs79" podUID="c8249efc-c3ac-4527-8394-e6e0aff9d457" containerName="registry-server" containerID="cri-o://82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae" gracePeriod=30 Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.029102 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m6g58"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.029348 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" podUID="46615e45-27d1-47b4-a5d7-f0cfb9f7476d" containerName="marketplace-operator" containerID="cri-o://112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132" gracePeriod=30 Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.040861 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6dlz8"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.041095 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6dlz8" podUID="32b94450-1b99-4fc9-aac8-658628a2195f" containerName="registry-server" containerID="cri-o://eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e" gracePeriod=30 Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.050236 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkfq7"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.051108 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.053568 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9q6bg"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.054033 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9q6bg" podUID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerName="registry-server" containerID="cri-o://8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c" gracePeriod=30 Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.069446 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkfq7"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.169216 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2efbcb2c-3e38-4eda-bf92-9805f287bbb3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wkfq7\" (UID: \"2efbcb2c-3e38-4eda-bf92-9805f287bbb3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.169271 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2efbcb2c-3e38-4eda-bf92-9805f287bbb3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wkfq7\" (UID: \"2efbcb2c-3e38-4eda-bf92-9805f287bbb3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.169309 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntsww\" (UniqueName: \"kubernetes.io/projected/2efbcb2c-3e38-4eda-bf92-9805f287bbb3-kube-api-access-ntsww\") pod \"marketplace-operator-79b997595-wkfq7\" (UID: \"2efbcb2c-3e38-4eda-bf92-9805f287bbb3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.271101 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2efbcb2c-3e38-4eda-bf92-9805f287bbb3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wkfq7\" (UID: \"2efbcb2c-3e38-4eda-bf92-9805f287bbb3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.271163 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2efbcb2c-3e38-4eda-bf92-9805f287bbb3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wkfq7\" (UID: \"2efbcb2c-3e38-4eda-bf92-9805f287bbb3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.271215 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntsww\" (UniqueName: \"kubernetes.io/projected/2efbcb2c-3e38-4eda-bf92-9805f287bbb3-kube-api-access-ntsww\") pod \"marketplace-operator-79b997595-wkfq7\" (UID: \"2efbcb2c-3e38-4eda-bf92-9805f287bbb3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.280375 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2efbcb2c-3e38-4eda-bf92-9805f287bbb3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wkfq7\" (UID: \"2efbcb2c-3e38-4eda-bf92-9805f287bbb3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.281425 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2efbcb2c-3e38-4eda-bf92-9805f287bbb3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wkfq7\" (UID: \"2efbcb2c-3e38-4eda-bf92-9805f287bbb3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.289964 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntsww\" (UniqueName: \"kubernetes.io/projected/2efbcb2c-3e38-4eda-bf92-9805f287bbb3-kube-api-access-ntsww\") pod \"marketplace-operator-79b997595-wkfq7\" (UID: \"2efbcb2c-3e38-4eda-bf92-9805f287bbb3\") " pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.418204 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.430734 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.579596 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlj4f\" (UniqueName: \"kubernetes.io/projected/9f532fb4-a915-4cf7-a64a-887eadfb1710-kube-api-access-hlj4f\") pod \"9f532fb4-a915-4cf7-a64a-887eadfb1710\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.579771 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-catalog-content\") pod \"9f532fb4-a915-4cf7-a64a-887eadfb1710\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.579826 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-utilities\") pod \"9f532fb4-a915-4cf7-a64a-887eadfb1710\" (UID: \"9f532fb4-a915-4cf7-a64a-887eadfb1710\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.583570 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.589785 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f532fb4-a915-4cf7-a64a-887eadfb1710-kube-api-access-hlj4f" (OuterVolumeSpecName: "kube-api-access-hlj4f") pod "9f532fb4-a915-4cf7-a64a-887eadfb1710" (UID: "9f532fb4-a915-4cf7-a64a-887eadfb1710"). InnerVolumeSpecName "kube-api-access-hlj4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.591218 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-utilities" (OuterVolumeSpecName: "utilities") pod "9f532fb4-a915-4cf7-a64a-887eadfb1710" (UID: "9f532fb4-a915-4cf7-a64a-887eadfb1710"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.598430 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.623628 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.625168 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.679886 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f532fb4-a915-4cf7-a64a-887eadfb1710" (UID: "9f532fb4-a915-4cf7-a64a-887eadfb1710"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.682421 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-operator-metrics\") pod \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.682494 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj9bd\" (UniqueName: \"kubernetes.io/projected/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-kube-api-access-cj9bd\") pod \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.682558 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-trusted-ca\") pod \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\" (UID: \"46615e45-27d1-47b4-a5d7-f0cfb9f7476d\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.682810 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.682832 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f532fb4-a915-4cf7-a64a-887eadfb1710-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.682843 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlj4f\" (UniqueName: \"kubernetes.io/projected/9f532fb4-a915-4cf7-a64a-887eadfb1710-kube-api-access-hlj4f\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.687610 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "46615e45-27d1-47b4-a5d7-f0cfb9f7476d" (UID: "46615e45-27d1-47b4-a5d7-f0cfb9f7476d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.688979 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "46615e45-27d1-47b4-a5d7-f0cfb9f7476d" (UID: "46615e45-27d1-47b4-a5d7-f0cfb9f7476d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.694985 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-kube-api-access-cj9bd" (OuterVolumeSpecName: "kube-api-access-cj9bd") pod "46615e45-27d1-47b4-a5d7-f0cfb9f7476d" (UID: "46615e45-27d1-47b4-a5d7-f0cfb9f7476d"). InnerVolumeSpecName "kube-api-access-cj9bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.728449 4944 generic.go:334] "Generic (PLEG): container finished" podID="9f532fb4-a915-4cf7-a64a-887eadfb1710" containerID="481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076" exitCode=0 Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.728508 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sphtd" event={"ID":"9f532fb4-a915-4cf7-a64a-887eadfb1710","Type":"ContainerDied","Data":"481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076"} Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.728534 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sphtd" event={"ID":"9f532fb4-a915-4cf7-a64a-887eadfb1710","Type":"ContainerDied","Data":"30d9b2775ddd90a1aec2a749f5f5fffb9ac403b0c9cbfdeba22eda7a27bbd80e"} Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.728549 4944 scope.go:117] "RemoveContainer" containerID="481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.728711 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sphtd" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.740002 4944 generic.go:334] "Generic (PLEG): container finished" podID="32b94450-1b99-4fc9-aac8-658628a2195f" containerID="eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e" exitCode=0 Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.740176 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6dlz8" event={"ID":"32b94450-1b99-4fc9-aac8-658628a2195f","Type":"ContainerDied","Data":"eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e"} Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.740209 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6dlz8" event={"ID":"32b94450-1b99-4fc9-aac8-658628a2195f","Type":"ContainerDied","Data":"f20602614f18845f7da5fdbf6e3ff1c37310e2ce58d6484cdd31ac5ccd019213"} Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.740334 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6dlz8" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.744809 4944 generic.go:334] "Generic (PLEG): container finished" podID="46615e45-27d1-47b4-a5d7-f0cfb9f7476d" containerID="112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132" exitCode=0 Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.744848 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.744888 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" event={"ID":"46615e45-27d1-47b4-a5d7-f0cfb9f7476d","Type":"ContainerDied","Data":"112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132"} Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.744913 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m6g58" event={"ID":"46615e45-27d1-47b4-a5d7-f0cfb9f7476d","Type":"ContainerDied","Data":"76c9207008b198cb2271af30032a9a9e2dd8d143bf18ec75883776947daccddf"} Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.775879 4944 scope.go:117] "RemoveContainer" containerID="d0addb826471836eb3c9b89984301c9c815db0e7b72526cafdd9ac9c613e08f4" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784081 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5dwm\" (UniqueName: \"kubernetes.io/projected/32b94450-1b99-4fc9-aac8-658628a2195f-kube-api-access-l5dwm\") pod \"32b94450-1b99-4fc9-aac8-658628a2195f\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784163 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-utilities\") pod \"32b94450-1b99-4fc9-aac8-658628a2195f\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784228 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-catalog-content\") pod \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784288 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-catalog-content\") pod \"32b94450-1b99-4fc9-aac8-658628a2195f\" (UID: \"32b94450-1b99-4fc9-aac8-658628a2195f\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784360 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-utilities\") pod \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784389 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-utilities\") pod \"c8249efc-c3ac-4527-8394-e6e0aff9d457\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784415 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxnxt\" (UniqueName: \"kubernetes.io/projected/c8249efc-c3ac-4527-8394-e6e0aff9d457-kube-api-access-kxnxt\") pod \"c8249efc-c3ac-4527-8394-e6e0aff9d457\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784453 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-catalog-content\") pod \"c8249efc-c3ac-4527-8394-e6e0aff9d457\" (UID: \"c8249efc-c3ac-4527-8394-e6e0aff9d457\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784518 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txvbs\" (UniqueName: \"kubernetes.io/projected/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-kube-api-access-txvbs\") pod \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\" (UID: \"6a34571c-5dc6-4ee5-9d59-7aa43d325b71\") " Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784817 4944 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784850 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj9bd\" (UniqueName: \"kubernetes.io/projected/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-kube-api-access-cj9bd\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.784864 4944 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/46615e45-27d1-47b4-a5d7-f0cfb9f7476d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.785263 4944 generic.go:334] "Generic (PLEG): container finished" podID="c8249efc-c3ac-4527-8394-e6e0aff9d457" containerID="82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae" exitCode=0 Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.785356 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8cs79" event={"ID":"c8249efc-c3ac-4527-8394-e6e0aff9d457","Type":"ContainerDied","Data":"82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae"} Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.785385 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8cs79" event={"ID":"c8249efc-c3ac-4527-8394-e6e0aff9d457","Type":"ContainerDied","Data":"e316def9707a1c3e9f2032c1f56b141c0955ac2a85c06b800e6ddec1da3c1079"} Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.785996 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8cs79" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.786669 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-utilities" (OuterVolumeSpecName: "utilities") pod "c8249efc-c3ac-4527-8394-e6e0aff9d457" (UID: "c8249efc-c3ac-4527-8394-e6e0aff9d457"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.787216 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-utilities" (OuterVolumeSpecName: "utilities") pod "6a34571c-5dc6-4ee5-9d59-7aa43d325b71" (UID: "6a34571c-5dc6-4ee5-9d59-7aa43d325b71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.787971 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-utilities" (OuterVolumeSpecName: "utilities") pod "32b94450-1b99-4fc9-aac8-658628a2195f" (UID: "32b94450-1b99-4fc9-aac8-658628a2195f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.797627 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32b94450-1b99-4fc9-aac8-658628a2195f-kube-api-access-l5dwm" (OuterVolumeSpecName: "kube-api-access-l5dwm") pod "32b94450-1b99-4fc9-aac8-658628a2195f" (UID: "32b94450-1b99-4fc9-aac8-658628a2195f"). InnerVolumeSpecName "kube-api-access-l5dwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.803131 4944 generic.go:334] "Generic (PLEG): container finished" podID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerID="8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c" exitCode=0 Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.803179 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6bg" event={"ID":"6a34571c-5dc6-4ee5-9d59-7aa43d325b71","Type":"ContainerDied","Data":"8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c"} Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.803644 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9q6bg" event={"ID":"6a34571c-5dc6-4ee5-9d59-7aa43d325b71","Type":"ContainerDied","Data":"43f34f98ef1e015aa3ba221695da1f46205d00ce0a70af695b30367a555982b9"} Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.803723 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9q6bg" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.805871 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wkfq7"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.806997 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8249efc-c3ac-4527-8394-e6e0aff9d457-kube-api-access-kxnxt" (OuterVolumeSpecName: "kube-api-access-kxnxt") pod "c8249efc-c3ac-4527-8394-e6e0aff9d457" (UID: "c8249efc-c3ac-4527-8394-e6e0aff9d457"). InnerVolumeSpecName "kube-api-access-kxnxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.820937 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-kube-api-access-txvbs" (OuterVolumeSpecName: "kube-api-access-txvbs") pod "6a34571c-5dc6-4ee5-9d59-7aa43d325b71" (UID: "6a34571c-5dc6-4ee5-9d59-7aa43d325b71"). InnerVolumeSpecName "kube-api-access-txvbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.824989 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sphtd"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.829128 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sphtd"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.831171 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32b94450-1b99-4fc9-aac8-658628a2195f" (UID: "32b94450-1b99-4fc9-aac8-658628a2195f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.831263 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m6g58"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.834511 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m6g58"] Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.847018 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8249efc-c3ac-4527-8394-e6e0aff9d457" (UID: "c8249efc-c3ac-4527-8394-e6e0aff9d457"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.852403 4944 scope.go:117] "RemoveContainer" containerID="5615ab50d9653d288de46ce0dffbb16d51a4f75ec6475a0c60212214fc0c6224" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.869432 4944 scope.go:117] "RemoveContainer" containerID="481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076" Nov 24 08:56:30 crc kubenswrapper[4944]: E1124 08:56:30.869973 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076\": container with ID starting with 481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076 not found: ID does not exist" containerID="481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.870025 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076"} err="failed to get container status \"481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076\": rpc error: code = NotFound desc = could not find container \"481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076\": container with ID starting with 481e54c69b6a00549d3b127b88ca552b98d831ec96474d88f8f0f9b14bbfd076 not found: ID does not exist" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.870079 4944 scope.go:117] "RemoveContainer" containerID="d0addb826471836eb3c9b89984301c9c815db0e7b72526cafdd9ac9c613e08f4" Nov 24 08:56:30 crc kubenswrapper[4944]: E1124 08:56:30.870331 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0addb826471836eb3c9b89984301c9c815db0e7b72526cafdd9ac9c613e08f4\": container with ID starting with d0addb826471836eb3c9b89984301c9c815db0e7b72526cafdd9ac9c613e08f4 not found: ID does not exist" containerID="d0addb826471836eb3c9b89984301c9c815db0e7b72526cafdd9ac9c613e08f4" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.870359 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0addb826471836eb3c9b89984301c9c815db0e7b72526cafdd9ac9c613e08f4"} err="failed to get container status \"d0addb826471836eb3c9b89984301c9c815db0e7b72526cafdd9ac9c613e08f4\": rpc error: code = NotFound desc = could not find container \"d0addb826471836eb3c9b89984301c9c815db0e7b72526cafdd9ac9c613e08f4\": container with ID starting with d0addb826471836eb3c9b89984301c9c815db0e7b72526cafdd9ac9c613e08f4 not found: ID does not exist" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.870375 4944 scope.go:117] "RemoveContainer" containerID="5615ab50d9653d288de46ce0dffbb16d51a4f75ec6475a0c60212214fc0c6224" Nov 24 08:56:30 crc kubenswrapper[4944]: E1124 08:56:30.870646 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5615ab50d9653d288de46ce0dffbb16d51a4f75ec6475a0c60212214fc0c6224\": container with ID starting with 5615ab50d9653d288de46ce0dffbb16d51a4f75ec6475a0c60212214fc0c6224 not found: ID does not exist" containerID="5615ab50d9653d288de46ce0dffbb16d51a4f75ec6475a0c60212214fc0c6224" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.870669 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5615ab50d9653d288de46ce0dffbb16d51a4f75ec6475a0c60212214fc0c6224"} err="failed to get container status \"5615ab50d9653d288de46ce0dffbb16d51a4f75ec6475a0c60212214fc0c6224\": rpc error: code = NotFound desc = could not find container \"5615ab50d9653d288de46ce0dffbb16d51a4f75ec6475a0c60212214fc0c6224\": container with ID starting with 5615ab50d9653d288de46ce0dffbb16d51a4f75ec6475a0c60212214fc0c6224 not found: ID does not exist" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.870684 4944 scope.go:117] "RemoveContainer" containerID="eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.884025 4944 scope.go:117] "RemoveContainer" containerID="c09928912211f9e65062ca36ab486e44fdb1ca6b6834c4e1058927a7b0fbd4c1" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.885960 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.885985 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.885995 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxnxt\" (UniqueName: \"kubernetes.io/projected/c8249efc-c3ac-4527-8394-e6e0aff9d457-kube-api-access-kxnxt\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.886006 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8249efc-c3ac-4527-8394-e6e0aff9d457-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.886015 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txvbs\" (UniqueName: \"kubernetes.io/projected/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-kube-api-access-txvbs\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.886024 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5dwm\" (UniqueName: \"kubernetes.io/projected/32b94450-1b99-4fc9-aac8-658628a2195f-kube-api-access-l5dwm\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.886032 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.886039 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32b94450-1b99-4fc9-aac8-658628a2195f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.891009 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a34571c-5dc6-4ee5-9d59-7aa43d325b71" (UID: "6a34571c-5dc6-4ee5-9d59-7aa43d325b71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.913258 4944 scope.go:117] "RemoveContainer" containerID="8655e4f1a7c5135556bf82b8285747f1c58c7e3c90c8d25f69c2205e80f2c782" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.944329 4944 scope.go:117] "RemoveContainer" containerID="eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e" Nov 24 08:56:30 crc kubenswrapper[4944]: E1124 08:56:30.944800 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e\": container with ID starting with eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e not found: ID does not exist" containerID="eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.944823 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e"} err="failed to get container status \"eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e\": rpc error: code = NotFound desc = could not find container \"eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e\": container with ID starting with eb78bdfd54708a762fcbd98d7ae830fd138b90bb4e520b92d08ac1c4dd39d08e not found: ID does not exist" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.944843 4944 scope.go:117] "RemoveContainer" containerID="c09928912211f9e65062ca36ab486e44fdb1ca6b6834c4e1058927a7b0fbd4c1" Nov 24 08:56:30 crc kubenswrapper[4944]: E1124 08:56:30.945246 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c09928912211f9e65062ca36ab486e44fdb1ca6b6834c4e1058927a7b0fbd4c1\": container with ID starting with c09928912211f9e65062ca36ab486e44fdb1ca6b6834c4e1058927a7b0fbd4c1 not found: ID does not exist" containerID="c09928912211f9e65062ca36ab486e44fdb1ca6b6834c4e1058927a7b0fbd4c1" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.945267 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c09928912211f9e65062ca36ab486e44fdb1ca6b6834c4e1058927a7b0fbd4c1"} err="failed to get container status \"c09928912211f9e65062ca36ab486e44fdb1ca6b6834c4e1058927a7b0fbd4c1\": rpc error: code = NotFound desc = could not find container \"c09928912211f9e65062ca36ab486e44fdb1ca6b6834c4e1058927a7b0fbd4c1\": container with ID starting with c09928912211f9e65062ca36ab486e44fdb1ca6b6834c4e1058927a7b0fbd4c1 not found: ID does not exist" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.945280 4944 scope.go:117] "RemoveContainer" containerID="8655e4f1a7c5135556bf82b8285747f1c58c7e3c90c8d25f69c2205e80f2c782" Nov 24 08:56:30 crc kubenswrapper[4944]: E1124 08:56:30.945581 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8655e4f1a7c5135556bf82b8285747f1c58c7e3c90c8d25f69c2205e80f2c782\": container with ID starting with 8655e4f1a7c5135556bf82b8285747f1c58c7e3c90c8d25f69c2205e80f2c782 not found: ID does not exist" containerID="8655e4f1a7c5135556bf82b8285747f1c58c7e3c90c8d25f69c2205e80f2c782" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.945600 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8655e4f1a7c5135556bf82b8285747f1c58c7e3c90c8d25f69c2205e80f2c782"} err="failed to get container status \"8655e4f1a7c5135556bf82b8285747f1c58c7e3c90c8d25f69c2205e80f2c782\": rpc error: code = NotFound desc = could not find container \"8655e4f1a7c5135556bf82b8285747f1c58c7e3c90c8d25f69c2205e80f2c782\": container with ID starting with 8655e4f1a7c5135556bf82b8285747f1c58c7e3c90c8d25f69c2205e80f2c782 not found: ID does not exist" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.945614 4944 scope.go:117] "RemoveContainer" containerID="112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.960353 4944 scope.go:117] "RemoveContainer" containerID="112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132" Nov 24 08:56:30 crc kubenswrapper[4944]: E1124 08:56:30.960736 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132\": container with ID starting with 112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132 not found: ID does not exist" containerID="112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.960771 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132"} err="failed to get container status \"112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132\": rpc error: code = NotFound desc = could not find container \"112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132\": container with ID starting with 112fd4c3630a14b6ac4377156d755d425f8612a737804a2789ad60c92bb75132 not found: ID does not exist" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.960794 4944 scope.go:117] "RemoveContainer" containerID="82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.974412 4944 scope.go:117] "RemoveContainer" containerID="7aea9faba41863011206d50fe384a02b9c8cddd84d5214a3d9e8b76670217a78" Nov 24 08:56:30 crc kubenswrapper[4944]: I1124 08:56:30.987481 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a34571c-5dc6-4ee5-9d59-7aa43d325b71-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.012860 4944 scope.go:117] "RemoveContainer" containerID="beaf3a9957410846a72137dc26b28e5cae51d9ed97c7a96cb087a05358fdff02" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.040082 4944 scope.go:117] "RemoveContainer" containerID="82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae" Nov 24 08:56:31 crc kubenswrapper[4944]: E1124 08:56:31.041854 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae\": container with ID starting with 82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae not found: ID does not exist" containerID="82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.041891 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae"} err="failed to get container status \"82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae\": rpc error: code = NotFound desc = could not find container \"82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae\": container with ID starting with 82945c1d5a01424db720d98536939b1157671f7bdd30d9d20245d75483aedaae not found: ID does not exist" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.041914 4944 scope.go:117] "RemoveContainer" containerID="7aea9faba41863011206d50fe384a02b9c8cddd84d5214a3d9e8b76670217a78" Nov 24 08:56:31 crc kubenswrapper[4944]: E1124 08:56:31.042412 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7aea9faba41863011206d50fe384a02b9c8cddd84d5214a3d9e8b76670217a78\": container with ID starting with 7aea9faba41863011206d50fe384a02b9c8cddd84d5214a3d9e8b76670217a78 not found: ID does not exist" containerID="7aea9faba41863011206d50fe384a02b9c8cddd84d5214a3d9e8b76670217a78" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.042477 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aea9faba41863011206d50fe384a02b9c8cddd84d5214a3d9e8b76670217a78"} err="failed to get container status \"7aea9faba41863011206d50fe384a02b9c8cddd84d5214a3d9e8b76670217a78\": rpc error: code = NotFound desc = could not find container \"7aea9faba41863011206d50fe384a02b9c8cddd84d5214a3d9e8b76670217a78\": container with ID starting with 7aea9faba41863011206d50fe384a02b9c8cddd84d5214a3d9e8b76670217a78 not found: ID does not exist" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.042518 4944 scope.go:117] "RemoveContainer" containerID="beaf3a9957410846a72137dc26b28e5cae51d9ed97c7a96cb087a05358fdff02" Nov 24 08:56:31 crc kubenswrapper[4944]: E1124 08:56:31.043081 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beaf3a9957410846a72137dc26b28e5cae51d9ed97c7a96cb087a05358fdff02\": container with ID starting with beaf3a9957410846a72137dc26b28e5cae51d9ed97c7a96cb087a05358fdff02 not found: ID does not exist" containerID="beaf3a9957410846a72137dc26b28e5cae51d9ed97c7a96cb087a05358fdff02" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.043128 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beaf3a9957410846a72137dc26b28e5cae51d9ed97c7a96cb087a05358fdff02"} err="failed to get container status \"beaf3a9957410846a72137dc26b28e5cae51d9ed97c7a96cb087a05358fdff02\": rpc error: code = NotFound desc = could not find container \"beaf3a9957410846a72137dc26b28e5cae51d9ed97c7a96cb087a05358fdff02\": container with ID starting with beaf3a9957410846a72137dc26b28e5cae51d9ed97c7a96cb087a05358fdff02 not found: ID does not exist" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.043160 4944 scope.go:117] "RemoveContainer" containerID="8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.058232 4944 scope.go:117] "RemoveContainer" containerID="80af712ab259b6f36b3ba2450cd0ce1001e5a0be3251e2d614e5b7fff99be199" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.073023 4944 scope.go:117] "RemoveContainer" containerID="6ac9bb1fd81282b0a1c6bdb0d059a6a3560d3710c8a1d00a1a936867d8053db9" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.074971 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6dlz8"] Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.077367 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6dlz8"] Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.088879 4944 scope.go:117] "RemoveContainer" containerID="8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c" Nov 24 08:56:31 crc kubenswrapper[4944]: E1124 08:56:31.089418 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c\": container with ID starting with 8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c not found: ID does not exist" containerID="8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.089461 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c"} err="failed to get container status \"8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c\": rpc error: code = NotFound desc = could not find container \"8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c\": container with ID starting with 8ac4c553361381933c9018d62861b4eb04de37dd42456a63c7865680132d787c not found: ID does not exist" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.089490 4944 scope.go:117] "RemoveContainer" containerID="80af712ab259b6f36b3ba2450cd0ce1001e5a0be3251e2d614e5b7fff99be199" Nov 24 08:56:31 crc kubenswrapper[4944]: E1124 08:56:31.089913 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80af712ab259b6f36b3ba2450cd0ce1001e5a0be3251e2d614e5b7fff99be199\": container with ID starting with 80af712ab259b6f36b3ba2450cd0ce1001e5a0be3251e2d614e5b7fff99be199 not found: ID does not exist" containerID="80af712ab259b6f36b3ba2450cd0ce1001e5a0be3251e2d614e5b7fff99be199" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.089981 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80af712ab259b6f36b3ba2450cd0ce1001e5a0be3251e2d614e5b7fff99be199"} err="failed to get container status \"80af712ab259b6f36b3ba2450cd0ce1001e5a0be3251e2d614e5b7fff99be199\": rpc error: code = NotFound desc = could not find container \"80af712ab259b6f36b3ba2450cd0ce1001e5a0be3251e2d614e5b7fff99be199\": container with ID starting with 80af712ab259b6f36b3ba2450cd0ce1001e5a0be3251e2d614e5b7fff99be199 not found: ID does not exist" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.090081 4944 scope.go:117] "RemoveContainer" containerID="6ac9bb1fd81282b0a1c6bdb0d059a6a3560d3710c8a1d00a1a936867d8053db9" Nov 24 08:56:31 crc kubenswrapper[4944]: E1124 08:56:31.090499 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ac9bb1fd81282b0a1c6bdb0d059a6a3560d3710c8a1d00a1a936867d8053db9\": container with ID starting with 6ac9bb1fd81282b0a1c6bdb0d059a6a3560d3710c8a1d00a1a936867d8053db9 not found: ID does not exist" containerID="6ac9bb1fd81282b0a1c6bdb0d059a6a3560d3710c8a1d00a1a936867d8053db9" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.090527 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ac9bb1fd81282b0a1c6bdb0d059a6a3560d3710c8a1d00a1a936867d8053db9"} err="failed to get container status \"6ac9bb1fd81282b0a1c6bdb0d059a6a3560d3710c8a1d00a1a936867d8053db9\": rpc error: code = NotFound desc = could not find container \"6ac9bb1fd81282b0a1c6bdb0d059a6a3560d3710c8a1d00a1a936867d8053db9\": container with ID starting with 6ac9bb1fd81282b0a1c6bdb0d059a6a3560d3710c8a1d00a1a936867d8053db9 not found: ID does not exist" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.119517 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8cs79"] Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.125005 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8cs79"] Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.135896 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9q6bg"] Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.138098 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9q6bg"] Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.813257 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" event={"ID":"2efbcb2c-3e38-4eda-bf92-9805f287bbb3","Type":"ContainerStarted","Data":"a0cbc7daec4ebd18c4232a513917f6d54677bf9714878b70460d5fb84d7902ef"} Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.813301 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" event={"ID":"2efbcb2c-3e38-4eda-bf92-9805f287bbb3","Type":"ContainerStarted","Data":"9befa65d4e0c8874a06170bac5c1781422b66b18d6f09ab1d87c2832955bbf65"} Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.813531 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.816565 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" Nov 24 08:56:31 crc kubenswrapper[4944]: I1124 08:56:31.837835 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wkfq7" podStartSLOduration=1.83781149 podStartE2EDuration="1.83781149s" podCreationTimestamp="2025-11-24 08:56:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:56:31.8337712 +0000 UTC m=+252.368211662" watchObservedRunningTime="2025-11-24 08:56:31.83781149 +0000 UTC m=+252.372251972" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235105 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-czxzn"] Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235599 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerName="extract-content" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235613 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerName="extract-content" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235625 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46615e45-27d1-47b4-a5d7-f0cfb9f7476d" containerName="marketplace-operator" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235631 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="46615e45-27d1-47b4-a5d7-f0cfb9f7476d" containerName="marketplace-operator" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235638 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f532fb4-a915-4cf7-a64a-887eadfb1710" containerName="extract-content" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235645 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f532fb4-a915-4cf7-a64a-887eadfb1710" containerName="extract-content" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235655 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b94450-1b99-4fc9-aac8-658628a2195f" containerName="extract-content" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235660 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b94450-1b99-4fc9-aac8-658628a2195f" containerName="extract-content" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235668 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8249efc-c3ac-4527-8394-e6e0aff9d457" containerName="extract-content" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235674 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8249efc-c3ac-4527-8394-e6e0aff9d457" containerName="extract-content" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235680 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8249efc-c3ac-4527-8394-e6e0aff9d457" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235686 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8249efc-c3ac-4527-8394-e6e0aff9d457" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235695 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8249efc-c3ac-4527-8394-e6e0aff9d457" containerName="extract-utilities" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235702 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8249efc-c3ac-4527-8394-e6e0aff9d457" containerName="extract-utilities" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235708 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f532fb4-a915-4cf7-a64a-887eadfb1710" containerName="extract-utilities" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235713 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f532fb4-a915-4cf7-a64a-887eadfb1710" containerName="extract-utilities" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235721 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b94450-1b99-4fc9-aac8-658628a2195f" containerName="extract-utilities" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235728 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b94450-1b99-4fc9-aac8-658628a2195f" containerName="extract-utilities" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235735 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b94450-1b99-4fc9-aac8-658628a2195f" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235741 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b94450-1b99-4fc9-aac8-658628a2195f" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235747 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f532fb4-a915-4cf7-a64a-887eadfb1710" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235753 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f532fb4-a915-4cf7-a64a-887eadfb1710" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235760 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerName="extract-utilities" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235766 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerName="extract-utilities" Nov 24 08:56:32 crc kubenswrapper[4944]: E1124 08:56:32.235775 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.235781 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.236847 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="46615e45-27d1-47b4-a5d7-f0cfb9f7476d" containerName="marketplace-operator" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.236890 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.236911 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f532fb4-a915-4cf7-a64a-887eadfb1710" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.236927 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8249efc-c3ac-4527-8394-e6e0aff9d457" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.236942 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="32b94450-1b99-4fc9-aac8-658628a2195f" containerName="registry-server" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.238848 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.248282 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.262029 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-czxzn"] Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.283886 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32b94450-1b99-4fc9-aac8-658628a2195f" path="/var/lib/kubelet/pods/32b94450-1b99-4fc9-aac8-658628a2195f/volumes" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.284590 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46615e45-27d1-47b4-a5d7-f0cfb9f7476d" path="/var/lib/kubelet/pods/46615e45-27d1-47b4-a5d7-f0cfb9f7476d/volumes" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.286327 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a34571c-5dc6-4ee5-9d59-7aa43d325b71" path="/var/lib/kubelet/pods/6a34571c-5dc6-4ee5-9d59-7aa43d325b71/volumes" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.287345 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f532fb4-a915-4cf7-a64a-887eadfb1710" path="/var/lib/kubelet/pods/9f532fb4-a915-4cf7-a64a-887eadfb1710/volumes" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.287914 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8249efc-c3ac-4527-8394-e6e0aff9d457" path="/var/lib/kubelet/pods/c8249efc-c3ac-4527-8394-e6e0aff9d457/volumes" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.403775 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ffe67e3-61b1-401f-af45-af86b8206f85-utilities\") pod \"redhat-marketplace-czxzn\" (UID: \"2ffe67e3-61b1-401f-af45-af86b8206f85\") " pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.403841 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ffe67e3-61b1-401f-af45-af86b8206f85-catalog-content\") pod \"redhat-marketplace-czxzn\" (UID: \"2ffe67e3-61b1-401f-af45-af86b8206f85\") " pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.403895 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5jwg\" (UniqueName: \"kubernetes.io/projected/2ffe67e3-61b1-401f-af45-af86b8206f85-kube-api-access-w5jwg\") pod \"redhat-marketplace-czxzn\" (UID: \"2ffe67e3-61b1-401f-af45-af86b8206f85\") " pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.443405 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kq7m4"] Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.444619 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.446174 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kq7m4"] Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.446935 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.504905 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5jwg\" (UniqueName: \"kubernetes.io/projected/2ffe67e3-61b1-401f-af45-af86b8206f85-kube-api-access-w5jwg\") pod \"redhat-marketplace-czxzn\" (UID: \"2ffe67e3-61b1-401f-af45-af86b8206f85\") " pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.505005 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ffe67e3-61b1-401f-af45-af86b8206f85-utilities\") pod \"redhat-marketplace-czxzn\" (UID: \"2ffe67e3-61b1-401f-af45-af86b8206f85\") " pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.505040 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ffe67e3-61b1-401f-af45-af86b8206f85-catalog-content\") pod \"redhat-marketplace-czxzn\" (UID: \"2ffe67e3-61b1-401f-af45-af86b8206f85\") " pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.505451 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ffe67e3-61b1-401f-af45-af86b8206f85-utilities\") pod \"redhat-marketplace-czxzn\" (UID: \"2ffe67e3-61b1-401f-af45-af86b8206f85\") " pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.505545 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ffe67e3-61b1-401f-af45-af86b8206f85-catalog-content\") pod \"redhat-marketplace-czxzn\" (UID: \"2ffe67e3-61b1-401f-af45-af86b8206f85\") " pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.524557 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5jwg\" (UniqueName: \"kubernetes.io/projected/2ffe67e3-61b1-401f-af45-af86b8206f85-kube-api-access-w5jwg\") pod \"redhat-marketplace-czxzn\" (UID: \"2ffe67e3-61b1-401f-af45-af86b8206f85\") " pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.566290 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.605977 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-catalog-content\") pod \"redhat-operators-kq7m4\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.606238 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-utilities\") pod \"redhat-operators-kq7m4\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.606358 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xw9x\" (UniqueName: \"kubernetes.io/projected/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-kube-api-access-9xw9x\") pod \"redhat-operators-kq7m4\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.707488 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-catalog-content\") pod \"redhat-operators-kq7m4\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.707891 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-utilities\") pod \"redhat-operators-kq7m4\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.707921 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xw9x\" (UniqueName: \"kubernetes.io/projected/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-kube-api-access-9xw9x\") pod \"redhat-operators-kq7m4\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.708576 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-catalog-content\") pod \"redhat-operators-kq7m4\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.708627 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-utilities\") pod \"redhat-operators-kq7m4\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.725587 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xw9x\" (UniqueName: \"kubernetes.io/projected/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-kube-api-access-9xw9x\") pod \"redhat-operators-kq7m4\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.751283 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-czxzn"] Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.765373 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:32 crc kubenswrapper[4944]: I1124 08:56:32.821327 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czxzn" event={"ID":"2ffe67e3-61b1-401f-af45-af86b8206f85","Type":"ContainerStarted","Data":"dd1caa8a1e92d29c7911dcbe433c0dea4fdbbe2337ea0f8594a3fbcdbfa2d641"} Nov 24 08:56:33 crc kubenswrapper[4944]: I1124 08:56:33.159013 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kq7m4"] Nov 24 08:56:33 crc kubenswrapper[4944]: W1124 08:56:33.165123 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a5adcc4_e667_43f1_b5d5_fdca62ceced7.slice/crio-17da6a16389fa2e0896a204ec4894b9d28e23a85ed87fed277cf75b79a665651 WatchSource:0}: Error finding container 17da6a16389fa2e0896a204ec4894b9d28e23a85ed87fed277cf75b79a665651: Status 404 returned error can't find the container with id 17da6a16389fa2e0896a204ec4894b9d28e23a85ed87fed277cf75b79a665651 Nov 24 08:56:33 crc kubenswrapper[4944]: I1124 08:56:33.826925 4944 generic.go:334] "Generic (PLEG): container finished" podID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" containerID="8f10f230ae51b931afaa7a5291e7ea55ddc33bf29364cc9278d5eb82b4957e3e" exitCode=0 Nov 24 08:56:33 crc kubenswrapper[4944]: I1124 08:56:33.827004 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7m4" event={"ID":"1a5adcc4-e667-43f1-b5d5-fdca62ceced7","Type":"ContainerDied","Data":"8f10f230ae51b931afaa7a5291e7ea55ddc33bf29364cc9278d5eb82b4957e3e"} Nov 24 08:56:33 crc kubenswrapper[4944]: I1124 08:56:33.827037 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7m4" event={"ID":"1a5adcc4-e667-43f1-b5d5-fdca62ceced7","Type":"ContainerStarted","Data":"17da6a16389fa2e0896a204ec4894b9d28e23a85ed87fed277cf75b79a665651"} Nov 24 08:56:33 crc kubenswrapper[4944]: I1124 08:56:33.828574 4944 generic.go:334] "Generic (PLEG): container finished" podID="2ffe67e3-61b1-401f-af45-af86b8206f85" containerID="af31fb1b8732aab7971795c39a315c2a0e1c6848231bc226edbe2ac044085f5a" exitCode=0 Nov 24 08:56:33 crc kubenswrapper[4944]: I1124 08:56:33.828999 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czxzn" event={"ID":"2ffe67e3-61b1-401f-af45-af86b8206f85","Type":"ContainerDied","Data":"af31fb1b8732aab7971795c39a315c2a0e1c6848231bc226edbe2ac044085f5a"} Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.636949 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4d962"] Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.638223 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.640440 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.684000 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4d962"] Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.732036 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e-utilities\") pod \"community-operators-4d962\" (UID: \"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e\") " pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.732136 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e-catalog-content\") pod \"community-operators-4d962\" (UID: \"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e\") " pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.732316 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wqht\" (UniqueName: \"kubernetes.io/projected/c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e-kube-api-access-2wqht\") pod \"community-operators-4d962\" (UID: \"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e\") " pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.833524 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wqht\" (UniqueName: \"kubernetes.io/projected/c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e-kube-api-access-2wqht\") pod \"community-operators-4d962\" (UID: \"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e\") " pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.833596 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e-utilities\") pod \"community-operators-4d962\" (UID: \"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e\") " pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.833641 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e-catalog-content\") pod \"community-operators-4d962\" (UID: \"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e\") " pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.834134 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e-catalog-content\") pod \"community-operators-4d962\" (UID: \"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e\") " pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.834682 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e-utilities\") pod \"community-operators-4d962\" (UID: \"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e\") " pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.846035 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wwf7x"] Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.847300 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.849758 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7m4" event={"ID":"1a5adcc4-e667-43f1-b5d5-fdca62ceced7","Type":"ContainerStarted","Data":"19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec"} Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.849896 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wwf7x"] Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.849907 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.862462 4944 generic.go:334] "Generic (PLEG): container finished" podID="2ffe67e3-61b1-401f-af45-af86b8206f85" containerID="82966bf987aca56d6b5d45400a958f11504c1eafce5519027084293ff10cb97e" exitCode=0 Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.862519 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czxzn" event={"ID":"2ffe67e3-61b1-401f-af45-af86b8206f85","Type":"ContainerDied","Data":"82966bf987aca56d6b5d45400a958f11504c1eafce5519027084293ff10cb97e"} Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.866567 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wqht\" (UniqueName: \"kubernetes.io/projected/c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e-kube-api-access-2wqht\") pod \"community-operators-4d962\" (UID: \"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e\") " pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.934683 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b0535ee-607f-481f-a60b-40909235fa96-catalog-content\") pod \"certified-operators-wwf7x\" (UID: \"6b0535ee-607f-481f-a60b-40909235fa96\") " pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.934805 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b0535ee-607f-481f-a60b-40909235fa96-utilities\") pod \"certified-operators-wwf7x\" (UID: \"6b0535ee-607f-481f-a60b-40909235fa96\") " pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:34 crc kubenswrapper[4944]: I1124 08:56:34.934828 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75g8w\" (UniqueName: \"kubernetes.io/projected/6b0535ee-607f-481f-a60b-40909235fa96-kube-api-access-75g8w\") pod \"certified-operators-wwf7x\" (UID: \"6b0535ee-607f-481f-a60b-40909235fa96\") " pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.000937 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.037343 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b0535ee-607f-481f-a60b-40909235fa96-utilities\") pod \"certified-operators-wwf7x\" (UID: \"6b0535ee-607f-481f-a60b-40909235fa96\") " pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.037807 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75g8w\" (UniqueName: \"kubernetes.io/projected/6b0535ee-607f-481f-a60b-40909235fa96-kube-api-access-75g8w\") pod \"certified-operators-wwf7x\" (UID: \"6b0535ee-607f-481f-a60b-40909235fa96\") " pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.037896 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b0535ee-607f-481f-a60b-40909235fa96-catalog-content\") pod \"certified-operators-wwf7x\" (UID: \"6b0535ee-607f-481f-a60b-40909235fa96\") " pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.038407 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b0535ee-607f-481f-a60b-40909235fa96-catalog-content\") pod \"certified-operators-wwf7x\" (UID: \"6b0535ee-607f-481f-a60b-40909235fa96\") " pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.038629 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b0535ee-607f-481f-a60b-40909235fa96-utilities\") pod \"certified-operators-wwf7x\" (UID: \"6b0535ee-607f-481f-a60b-40909235fa96\") " pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.054830 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75g8w\" (UniqueName: \"kubernetes.io/projected/6b0535ee-607f-481f-a60b-40909235fa96-kube-api-access-75g8w\") pod \"certified-operators-wwf7x\" (UID: \"6b0535ee-607f-481f-a60b-40909235fa96\") " pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.170904 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.340596 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wwf7x"] Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.386304 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4d962"] Nov 24 08:56:35 crc kubenswrapper[4944]: W1124 08:56:35.393367 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8b3137f_dfa3_452a_a88e_4b08b6cfaa7e.slice/crio-8aec268e930db73237290ff847da96667763c5c118c52f9bc0d0477c445d2cce WatchSource:0}: Error finding container 8aec268e930db73237290ff847da96667763c5c118c52f9bc0d0477c445d2cce: Status 404 returned error can't find the container with id 8aec268e930db73237290ff847da96667763c5c118c52f9bc0d0477c445d2cce Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.869969 4944 generic.go:334] "Generic (PLEG): container finished" podID="c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e" containerID="de7fb1819beb5f5614a189d4aee0017c8293347c9db671b09f693cc70163e2a4" exitCode=0 Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.870082 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d962" event={"ID":"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e","Type":"ContainerDied","Data":"de7fb1819beb5f5614a189d4aee0017c8293347c9db671b09f693cc70163e2a4"} Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.870390 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d962" event={"ID":"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e","Type":"ContainerStarted","Data":"8aec268e930db73237290ff847da96667763c5c118c52f9bc0d0477c445d2cce"} Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.872962 4944 generic.go:334] "Generic (PLEG): container finished" podID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" containerID="19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec" exitCode=0 Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.873008 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7m4" event={"ID":"1a5adcc4-e667-43f1-b5d5-fdca62ceced7","Type":"ContainerDied","Data":"19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec"} Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.875505 4944 generic.go:334] "Generic (PLEG): container finished" podID="6b0535ee-607f-481f-a60b-40909235fa96" containerID="6585a7b5c82d56e42b44efaf9eaab3c2f7ab11d3421c4ca6291d10f2aa164170" exitCode=0 Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.875550 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wwf7x" event={"ID":"6b0535ee-607f-481f-a60b-40909235fa96","Type":"ContainerDied","Data":"6585a7b5c82d56e42b44efaf9eaab3c2f7ab11d3421c4ca6291d10f2aa164170"} Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.875573 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wwf7x" event={"ID":"6b0535ee-607f-481f-a60b-40909235fa96","Type":"ContainerStarted","Data":"982bf70888a7e44512f1412c17387342e2c6bc61da727aa124f7e10f8c7a4eb6"} Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.879695 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czxzn" event={"ID":"2ffe67e3-61b1-401f-af45-af86b8206f85","Type":"ContainerStarted","Data":"32286d30880fd7b7e2f6ea7234f05c684e6557b82b6ab10579b5beee6e0b21ad"} Nov 24 08:56:35 crc kubenswrapper[4944]: I1124 08:56:35.901828 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-czxzn" podStartSLOduration=2.333740079 podStartE2EDuration="3.901807116s" podCreationTimestamp="2025-11-24 08:56:32 +0000 UTC" firstStartedPulling="2025-11-24 08:56:33.830072495 +0000 UTC m=+254.364512967" lastFinishedPulling="2025-11-24 08:56:35.398139542 +0000 UTC m=+255.932580004" observedRunningTime="2025-11-24 08:56:35.901534537 +0000 UTC m=+256.435974999" watchObservedRunningTime="2025-11-24 08:56:35.901807116 +0000 UTC m=+256.436247588" Nov 24 08:56:36 crc kubenswrapper[4944]: I1124 08:56:36.890653 4944 generic.go:334] "Generic (PLEG): container finished" podID="c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e" containerID="e2ece7597d72aceadbaef4c77955da269b30e3208bd3ef5b6adbb6847098f24e" exitCode=0 Nov 24 08:56:36 crc kubenswrapper[4944]: I1124 08:56:36.890806 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d962" event={"ID":"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e","Type":"ContainerDied","Data":"e2ece7597d72aceadbaef4c77955da269b30e3208bd3ef5b6adbb6847098f24e"} Nov 24 08:56:36 crc kubenswrapper[4944]: I1124 08:56:36.894624 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7m4" event={"ID":"1a5adcc4-e667-43f1-b5d5-fdca62ceced7","Type":"ContainerStarted","Data":"c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0"} Nov 24 08:56:36 crc kubenswrapper[4944]: I1124 08:56:36.898388 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wwf7x" event={"ID":"6b0535ee-607f-481f-a60b-40909235fa96","Type":"ContainerStarted","Data":"a4d507c615496704318c411e592b5a7ccf0cd2156a53b24c2acf1b97f524537f"} Nov 24 08:56:36 crc kubenswrapper[4944]: I1124 08:56:36.953887 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kq7m4" podStartSLOduration=2.414010539 podStartE2EDuration="4.953863471s" podCreationTimestamp="2025-11-24 08:56:32 +0000 UTC" firstStartedPulling="2025-11-24 08:56:33.828973819 +0000 UTC m=+254.363414281" lastFinishedPulling="2025-11-24 08:56:36.368826751 +0000 UTC m=+256.903267213" observedRunningTime="2025-11-24 08:56:36.937919789 +0000 UTC m=+257.472360261" watchObservedRunningTime="2025-11-24 08:56:36.953863471 +0000 UTC m=+257.488303933" Nov 24 08:56:37 crc kubenswrapper[4944]: I1124 08:56:37.904137 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4d962" event={"ID":"c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e","Type":"ContainerStarted","Data":"dd8bb3fe67d78f9dac3dba2492d5421cb336ee2b40148533023911d66e5d714f"} Nov 24 08:56:37 crc kubenswrapper[4944]: I1124 08:56:37.906727 4944 generic.go:334] "Generic (PLEG): container finished" podID="6b0535ee-607f-481f-a60b-40909235fa96" containerID="a4d507c615496704318c411e592b5a7ccf0cd2156a53b24c2acf1b97f524537f" exitCode=0 Nov 24 08:56:37 crc kubenswrapper[4944]: I1124 08:56:37.906760 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wwf7x" event={"ID":"6b0535ee-607f-481f-a60b-40909235fa96","Type":"ContainerDied","Data":"a4d507c615496704318c411e592b5a7ccf0cd2156a53b24c2acf1b97f524537f"} Nov 24 08:56:37 crc kubenswrapper[4944]: I1124 08:56:37.926339 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4d962" podStartSLOduration=2.463664841 podStartE2EDuration="3.926319238s" podCreationTimestamp="2025-11-24 08:56:34 +0000 UTC" firstStartedPulling="2025-11-24 08:56:35.871682957 +0000 UTC m=+256.406123419" lastFinishedPulling="2025-11-24 08:56:37.334337354 +0000 UTC m=+257.868777816" observedRunningTime="2025-11-24 08:56:37.923925771 +0000 UTC m=+258.458366233" watchObservedRunningTime="2025-11-24 08:56:37.926319238 +0000 UTC m=+258.460759700" Nov 24 08:56:39 crc kubenswrapper[4944]: I1124 08:56:39.922824 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wwf7x" event={"ID":"6b0535ee-607f-481f-a60b-40909235fa96","Type":"ContainerStarted","Data":"f6f789f7a2cac67cfb4b9ef7d895b91e6de9654dc9c272f3ad4468052a98b126"} Nov 24 08:56:42 crc kubenswrapper[4944]: I1124 08:56:42.567298 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:42 crc kubenswrapper[4944]: I1124 08:56:42.567934 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:42 crc kubenswrapper[4944]: I1124 08:56:42.610318 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:42 crc kubenswrapper[4944]: I1124 08:56:42.629409 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wwf7x" podStartSLOduration=6.193674029 podStartE2EDuration="8.629393432s" podCreationTimestamp="2025-11-24 08:56:34 +0000 UTC" firstStartedPulling="2025-11-24 08:56:35.87765985 +0000 UTC m=+256.412100312" lastFinishedPulling="2025-11-24 08:56:38.313379253 +0000 UTC m=+258.847819715" observedRunningTime="2025-11-24 08:56:39.949649872 +0000 UTC m=+260.484090334" watchObservedRunningTime="2025-11-24 08:56:42.629393432 +0000 UTC m=+263.163833894" Nov 24 08:56:42 crc kubenswrapper[4944]: I1124 08:56:42.766499 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:42 crc kubenswrapper[4944]: I1124 08:56:42.766597 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:42 crc kubenswrapper[4944]: I1124 08:56:42.801581 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:42 crc kubenswrapper[4944]: I1124 08:56:42.978652 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 08:56:42 crc kubenswrapper[4944]: I1124 08:56:42.982705 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-czxzn" Nov 24 08:56:45 crc kubenswrapper[4944]: I1124 08:56:45.002315 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:45 crc kubenswrapper[4944]: I1124 08:56:45.002643 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:45 crc kubenswrapper[4944]: I1124 08:56:45.041707 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4d962" Nov 24 08:56:45 crc kubenswrapper[4944]: I1124 08:56:45.172056 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:45 crc kubenswrapper[4944]: I1124 08:56:45.172106 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:45 crc kubenswrapper[4944]: I1124 08:56:45.210623 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:45 crc kubenswrapper[4944]: I1124 08:56:45.998274 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wwf7x" Nov 24 08:56:45 crc kubenswrapper[4944]: I1124 08:56:45.999596 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4d962" Nov 24 08:57:53 crc kubenswrapper[4944]: I1124 08:57:53.548651 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:57:53 crc kubenswrapper[4944]: I1124 08:57:53.549455 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:58:23 crc kubenswrapper[4944]: I1124 08:58:23.549019 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:58:23 crc kubenswrapper[4944]: I1124 08:58:23.549603 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:58:53 crc kubenswrapper[4944]: I1124 08:58:53.549353 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:58:53 crc kubenswrapper[4944]: I1124 08:58:53.550317 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:58:53 crc kubenswrapper[4944]: I1124 08:58:53.550389 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 08:58:53 crc kubenswrapper[4944]: I1124 08:58:53.551404 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"152ff642f70a7f3a79f32127c3fc1fd67cb513fd122cd4fb960ccda5b2c27d78"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:58:53 crc kubenswrapper[4944]: I1124 08:58:53.551474 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://152ff642f70a7f3a79f32127c3fc1fd67cb513fd122cd4fb960ccda5b2c27d78" gracePeriod=600 Nov 24 08:58:54 crc kubenswrapper[4944]: I1124 08:58:54.594158 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="152ff642f70a7f3a79f32127c3fc1fd67cb513fd122cd4fb960ccda5b2c27d78" exitCode=0 Nov 24 08:58:54 crc kubenswrapper[4944]: I1124 08:58:54.594257 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"152ff642f70a7f3a79f32127c3fc1fd67cb513fd122cd4fb960ccda5b2c27d78"} Nov 24 08:58:54 crc kubenswrapper[4944]: I1124 08:58:54.595081 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"f05b7d2e929ec321cc8c8fd2068636a82f71fce547e3746a863561b7880eca47"} Nov 24 08:58:54 crc kubenswrapper[4944]: I1124 08:58:54.595145 4944 scope.go:117] "RemoveContainer" containerID="33729728882e1eacea01b781de4cf4207287b7a6c27a08030ab4d7a117699e83" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.134981 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd"] Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.136496 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.138765 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.139416 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.151426 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd"] Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.331739 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac60169b-7886-4307-abca-27316f3c4d0f-config-volume\") pod \"collect-profiles-29399580-bwtnd\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.331791 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac60169b-7886-4307-abca-27316f3c4d0f-secret-volume\") pod \"collect-profiles-29399580-bwtnd\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.331812 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clvrn\" (UniqueName: \"kubernetes.io/projected/ac60169b-7886-4307-abca-27316f3c4d0f-kube-api-access-clvrn\") pod \"collect-profiles-29399580-bwtnd\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.433284 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac60169b-7886-4307-abca-27316f3c4d0f-config-volume\") pod \"collect-profiles-29399580-bwtnd\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.433331 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac60169b-7886-4307-abca-27316f3c4d0f-secret-volume\") pod \"collect-profiles-29399580-bwtnd\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.433363 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clvrn\" (UniqueName: \"kubernetes.io/projected/ac60169b-7886-4307-abca-27316f3c4d0f-kube-api-access-clvrn\") pod \"collect-profiles-29399580-bwtnd\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.434326 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac60169b-7886-4307-abca-27316f3c4d0f-config-volume\") pod \"collect-profiles-29399580-bwtnd\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.444628 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac60169b-7886-4307-abca-27316f3c4d0f-secret-volume\") pod \"collect-profiles-29399580-bwtnd\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.453022 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clvrn\" (UniqueName: \"kubernetes.io/projected/ac60169b-7886-4307-abca-27316f3c4d0f-kube-api-access-clvrn\") pod \"collect-profiles-29399580-bwtnd\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.460759 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.693403 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd"] Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.988535 4944 generic.go:334] "Generic (PLEG): container finished" podID="ac60169b-7886-4307-abca-27316f3c4d0f" containerID="01a8d99f749f4ef70f2e5f5c87c64e59aa075551a45050e3ef9750c374dd16b1" exitCode=0 Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.988608 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" event={"ID":"ac60169b-7886-4307-abca-27316f3c4d0f","Type":"ContainerDied","Data":"01a8d99f749f4ef70f2e5f5c87c64e59aa075551a45050e3ef9750c374dd16b1"} Nov 24 09:00:00 crc kubenswrapper[4944]: I1124 09:00:00.988653 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" event={"ID":"ac60169b-7886-4307-abca-27316f3c4d0f","Type":"ContainerStarted","Data":"8407788381d3558db54524ba59353f8d3c4d765bd3dcd0d93f90c36dee1a2ce9"} Nov 24 09:00:02 crc kubenswrapper[4944]: I1124 09:00:02.184945 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:02 crc kubenswrapper[4944]: I1124 09:00:02.361362 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clvrn\" (UniqueName: \"kubernetes.io/projected/ac60169b-7886-4307-abca-27316f3c4d0f-kube-api-access-clvrn\") pod \"ac60169b-7886-4307-abca-27316f3c4d0f\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " Nov 24 09:00:02 crc kubenswrapper[4944]: I1124 09:00:02.361412 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac60169b-7886-4307-abca-27316f3c4d0f-config-volume\") pod \"ac60169b-7886-4307-abca-27316f3c4d0f\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " Nov 24 09:00:02 crc kubenswrapper[4944]: I1124 09:00:02.361461 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac60169b-7886-4307-abca-27316f3c4d0f-secret-volume\") pod \"ac60169b-7886-4307-abca-27316f3c4d0f\" (UID: \"ac60169b-7886-4307-abca-27316f3c4d0f\") " Nov 24 09:00:02 crc kubenswrapper[4944]: I1124 09:00:02.362130 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac60169b-7886-4307-abca-27316f3c4d0f-config-volume" (OuterVolumeSpecName: "config-volume") pod "ac60169b-7886-4307-abca-27316f3c4d0f" (UID: "ac60169b-7886-4307-abca-27316f3c4d0f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:00:02 crc kubenswrapper[4944]: I1124 09:00:02.366836 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac60169b-7886-4307-abca-27316f3c4d0f-kube-api-access-clvrn" (OuterVolumeSpecName: "kube-api-access-clvrn") pod "ac60169b-7886-4307-abca-27316f3c4d0f" (UID: "ac60169b-7886-4307-abca-27316f3c4d0f"). InnerVolumeSpecName "kube-api-access-clvrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:00:02 crc kubenswrapper[4944]: I1124 09:00:02.366979 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac60169b-7886-4307-abca-27316f3c4d0f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ac60169b-7886-4307-abca-27316f3c4d0f" (UID: "ac60169b-7886-4307-abca-27316f3c4d0f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:00:02 crc kubenswrapper[4944]: I1124 09:00:02.463158 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clvrn\" (UniqueName: \"kubernetes.io/projected/ac60169b-7886-4307-abca-27316f3c4d0f-kube-api-access-clvrn\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:02 crc kubenswrapper[4944]: I1124 09:00:02.463188 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac60169b-7886-4307-abca-27316f3c4d0f-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:02 crc kubenswrapper[4944]: I1124 09:00:02.463197 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac60169b-7886-4307-abca-27316f3c4d0f-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:03 crc kubenswrapper[4944]: I1124 09:00:03.002416 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" event={"ID":"ac60169b-7886-4307-abca-27316f3c4d0f","Type":"ContainerDied","Data":"8407788381d3558db54524ba59353f8d3c4d765bd3dcd0d93f90c36dee1a2ce9"} Nov 24 09:00:03 crc kubenswrapper[4944]: I1124 09:00:03.002728 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8407788381d3558db54524ba59353f8d3c4d765bd3dcd0d93f90c36dee1a2ce9" Nov 24 09:00:03 crc kubenswrapper[4944]: I1124 09:00:03.002470 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.753368 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-94rfs"] Nov 24 09:00:07 crc kubenswrapper[4944]: E1124 09:00:07.753898 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac60169b-7886-4307-abca-27316f3c4d0f" containerName="collect-profiles" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.753912 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac60169b-7886-4307-abca-27316f3c4d0f" containerName="collect-profiles" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.754014 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac60169b-7886-4307-abca-27316f3c4d0f" containerName="collect-profiles" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.754480 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.775125 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-94rfs"] Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.924239 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-ca-trust-extracted\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.924290 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-registry-tls\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.924314 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.924335 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfznf\" (UniqueName: \"kubernetes.io/projected/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-kube-api-access-bfznf\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.924494 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-installation-pull-secrets\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.924513 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-bound-sa-token\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.924533 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-registry-certificates\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.924554 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-trusted-ca\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:07 crc kubenswrapper[4944]: I1124 09:00:07.946349 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.025263 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-registry-tls\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.025304 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfznf\" (UniqueName: \"kubernetes.io/projected/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-kube-api-access-bfznf\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.025339 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-installation-pull-secrets\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.025358 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-bound-sa-token\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.025380 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-registry-certificates\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.025402 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-trusted-ca\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.025450 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-ca-trust-extracted\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.027142 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-registry-certificates\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.027283 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-trusted-ca\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.027394 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-ca-trust-extracted\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.030927 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-registry-tls\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.031036 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-installation-pull-secrets\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.041834 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfznf\" (UniqueName: \"kubernetes.io/projected/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-kube-api-access-bfznf\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.043067 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dbfe924-d3aa-4cdb-b05d-16caff15eb32-bound-sa-token\") pod \"image-registry-66df7c8f76-94rfs\" (UID: \"3dbfe924-d3aa-4cdb-b05d-16caff15eb32\") " pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.069021 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:08 crc kubenswrapper[4944]: I1124 09:00:08.233630 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-94rfs"] Nov 24 09:00:09 crc kubenswrapper[4944]: I1124 09:00:09.032711 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" event={"ID":"3dbfe924-d3aa-4cdb-b05d-16caff15eb32","Type":"ContainerStarted","Data":"5f51bec7680ef1086ee5f3ce6d9544a0906b929a2b8d1a0c928dec7c56d6f973"} Nov 24 09:00:09 crc kubenswrapper[4944]: I1124 09:00:09.032955 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" event={"ID":"3dbfe924-d3aa-4cdb-b05d-16caff15eb32","Type":"ContainerStarted","Data":"ac38e6f1d5c02efe9c122d2f330e5ed2ec127765c14ecc84f275c06d1e0449d0"} Nov 24 09:00:09 crc kubenswrapper[4944]: I1124 09:00:09.033073 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:09 crc kubenswrapper[4944]: I1124 09:00:09.055270 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" podStartSLOduration=2.055253071 podStartE2EDuration="2.055253071s" podCreationTimestamp="2025-11-24 09:00:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:00:09.053289088 +0000 UTC m=+469.587729550" watchObservedRunningTime="2025-11-24 09:00:09.055253071 +0000 UTC m=+469.589693533" Nov 24 09:00:28 crc kubenswrapper[4944]: I1124 09:00:28.074885 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-94rfs" Nov 24 09:00:28 crc kubenswrapper[4944]: I1124 09:00:28.120622 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v5288"] Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.158478 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" podUID="17c159e8-7de8-4716-b4cd-7f28a257a2bb" containerName="registry" containerID="cri-o://4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f" gracePeriod=30 Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.495541 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.549163 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.549253 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.610393 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-bound-sa-token\") pod \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.610469 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-certificates\") pod \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.610518 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/17c159e8-7de8-4716-b4cd-7f28a257a2bb-ca-trust-extracted\") pod \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.610562 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-trusted-ca\") pod \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.610588 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/17c159e8-7de8-4716-b4cd-7f28a257a2bb-installation-pull-secrets\") pod \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.610626 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4wgl\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-kube-api-access-g4wgl\") pod \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.610652 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-tls\") pod \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.610757 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\" (UID: \"17c159e8-7de8-4716-b4cd-7f28a257a2bb\") " Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.611383 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "17c159e8-7de8-4716-b4cd-7f28a257a2bb" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.611431 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "17c159e8-7de8-4716-b4cd-7f28a257a2bb" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.616703 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17c159e8-7de8-4716-b4cd-7f28a257a2bb-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "17c159e8-7de8-4716-b4cd-7f28a257a2bb" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.616778 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-kube-api-access-g4wgl" (OuterVolumeSpecName: "kube-api-access-g4wgl") pod "17c159e8-7de8-4716-b4cd-7f28a257a2bb" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb"). InnerVolumeSpecName "kube-api-access-g4wgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.616983 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "17c159e8-7de8-4716-b4cd-7f28a257a2bb" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.618056 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "17c159e8-7de8-4716-b4cd-7f28a257a2bb" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.627523 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17c159e8-7de8-4716-b4cd-7f28a257a2bb-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "17c159e8-7de8-4716-b4cd-7f28a257a2bb" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.633802 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "17c159e8-7de8-4716-b4cd-7f28a257a2bb" (UID: "17c159e8-7de8-4716-b4cd-7f28a257a2bb"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.712418 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4wgl\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-kube-api-access-g4wgl\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.712451 4944 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.712467 4944 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/17c159e8-7de8-4716-b4cd-7f28a257a2bb-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.712484 4944 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.712496 4944 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/17c159e8-7de8-4716-b4cd-7f28a257a2bb-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.712507 4944 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/17c159e8-7de8-4716-b4cd-7f28a257a2bb-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:53 crc kubenswrapper[4944]: I1124 09:00:53.712522 4944 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/17c159e8-7de8-4716-b4cd-7f28a257a2bb-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:54 crc kubenswrapper[4944]: I1124 09:00:54.249133 4944 generic.go:334] "Generic (PLEG): container finished" podID="17c159e8-7de8-4716-b4cd-7f28a257a2bb" containerID="4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f" exitCode=0 Nov 24 09:00:54 crc kubenswrapper[4944]: I1124 09:00:54.249206 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" event={"ID":"17c159e8-7de8-4716-b4cd-7f28a257a2bb","Type":"ContainerDied","Data":"4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f"} Nov 24 09:00:54 crc kubenswrapper[4944]: I1124 09:00:54.249449 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" event={"ID":"17c159e8-7de8-4716-b4cd-7f28a257a2bb","Type":"ContainerDied","Data":"96bda08e2027b94d28c583ee2b21129486167239153126cc50297909a911e15b"} Nov 24 09:00:54 crc kubenswrapper[4944]: I1124 09:00:54.249478 4944 scope.go:117] "RemoveContainer" containerID="4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f" Nov 24 09:00:54 crc kubenswrapper[4944]: I1124 09:00:54.249259 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-v5288" Nov 24 09:00:54 crc kubenswrapper[4944]: I1124 09:00:54.278636 4944 scope.go:117] "RemoveContainer" containerID="4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f" Nov 24 09:00:54 crc kubenswrapper[4944]: E1124 09:00:54.279349 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f\": container with ID starting with 4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f not found: ID does not exist" containerID="4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f" Nov 24 09:00:54 crc kubenswrapper[4944]: I1124 09:00:54.279382 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f"} err="failed to get container status \"4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f\": rpc error: code = NotFound desc = could not find container \"4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f\": container with ID starting with 4597b1efc27c33d42094e233771c8a0535451729e77fd9135e8cb9df59f0832f not found: ID does not exist" Nov 24 09:00:54 crc kubenswrapper[4944]: I1124 09:00:54.287446 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v5288"] Nov 24 09:00:54 crc kubenswrapper[4944]: I1124 09:00:54.287484 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-v5288"] Nov 24 09:00:56 crc kubenswrapper[4944]: I1124 09:00:56.283084 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17c159e8-7de8-4716-b4cd-7f28a257a2bb" path="/var/lib/kubelet/pods/17c159e8-7de8-4716-b4cd-7f28a257a2bb/volumes" Nov 24 09:01:23 crc kubenswrapper[4944]: I1124 09:01:23.548591 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:01:23 crc kubenswrapper[4944]: I1124 09:01:23.549203 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:01:53 crc kubenswrapper[4944]: I1124 09:01:53.549090 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:01:53 crc kubenswrapper[4944]: I1124 09:01:53.549533 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:01:53 crc kubenswrapper[4944]: I1124 09:01:53.549579 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:01:53 crc kubenswrapper[4944]: I1124 09:01:53.550182 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f05b7d2e929ec321cc8c8fd2068636a82f71fce547e3746a863561b7880eca47"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:01:53 crc kubenswrapper[4944]: I1124 09:01:53.550226 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://f05b7d2e929ec321cc8c8fd2068636a82f71fce547e3746a863561b7880eca47" gracePeriod=600 Nov 24 09:01:54 crc kubenswrapper[4944]: I1124 09:01:54.555342 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="f05b7d2e929ec321cc8c8fd2068636a82f71fce547e3746a863561b7880eca47" exitCode=0 Nov 24 09:01:54 crc kubenswrapper[4944]: I1124 09:01:54.555462 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"f05b7d2e929ec321cc8c8fd2068636a82f71fce547e3746a863561b7880eca47"} Nov 24 09:01:54 crc kubenswrapper[4944]: I1124 09:01:54.556072 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"f61d46fb34a7fb69e551ab18c2960366dce6b114b94216eab6db96b03be79c6f"} Nov 24 09:01:54 crc kubenswrapper[4944]: I1124 09:01:54.556117 4944 scope.go:117] "RemoveContainer" containerID="152ff642f70a7f3a79f32127c3fc1fd67cb513fd122cd4fb960ccda5b2c27d78" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.032691 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-462nz"] Nov 24 09:03:35 crc kubenswrapper[4944]: E1124 09:03:35.034116 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17c159e8-7de8-4716-b4cd-7f28a257a2bb" containerName="registry" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.034139 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="17c159e8-7de8-4716-b4cd-7f28a257a2bb" containerName="registry" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.034298 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="17c159e8-7de8-4716-b4cd-7f28a257a2bb" containerName="registry" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.034958 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.038758 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-462nz"] Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.079496 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.079828 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.079961 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.080126 4944 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-8fr8r" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.112869 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7372c865-b00f-4772-89c8-3e5a8a662eb6-crc-storage\") pod \"crc-storage-crc-462nz\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.113285 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w27vt\" (UniqueName: \"kubernetes.io/projected/7372c865-b00f-4772-89c8-3e5a8a662eb6-kube-api-access-w27vt\") pod \"crc-storage-crc-462nz\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.113370 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7372c865-b00f-4772-89c8-3e5a8a662eb6-node-mnt\") pod \"crc-storage-crc-462nz\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.214283 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7372c865-b00f-4772-89c8-3e5a8a662eb6-node-mnt\") pod \"crc-storage-crc-462nz\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.214372 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7372c865-b00f-4772-89c8-3e5a8a662eb6-crc-storage\") pod \"crc-storage-crc-462nz\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.214391 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w27vt\" (UniqueName: \"kubernetes.io/projected/7372c865-b00f-4772-89c8-3e5a8a662eb6-kube-api-access-w27vt\") pod \"crc-storage-crc-462nz\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.214592 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7372c865-b00f-4772-89c8-3e5a8a662eb6-node-mnt\") pod \"crc-storage-crc-462nz\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.215408 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7372c865-b00f-4772-89c8-3e5a8a662eb6-crc-storage\") pod \"crc-storage-crc-462nz\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.234973 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w27vt\" (UniqueName: \"kubernetes.io/projected/7372c865-b00f-4772-89c8-3e5a8a662eb6-kube-api-access-w27vt\") pod \"crc-storage-crc-462nz\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.395310 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.568930 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-462nz"] Nov 24 09:03:35 crc kubenswrapper[4944]: I1124 09:03:35.577683 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:03:36 crc kubenswrapper[4944]: I1124 09:03:36.093218 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-462nz" event={"ID":"7372c865-b00f-4772-89c8-3e5a8a662eb6","Type":"ContainerStarted","Data":"159b728deca143210b65f13ed48499156dd3307b636ea19873eedb3c5e7f6c27"} Nov 24 09:03:37 crc kubenswrapper[4944]: I1124 09:03:37.102944 4944 generic.go:334] "Generic (PLEG): container finished" podID="7372c865-b00f-4772-89c8-3e5a8a662eb6" containerID="9a15994acf814d96da9fbcb84987a2edeb10af89e1b301670246344f9659e481" exitCode=0 Nov 24 09:03:37 crc kubenswrapper[4944]: I1124 09:03:37.103019 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-462nz" event={"ID":"7372c865-b00f-4772-89c8-3e5a8a662eb6","Type":"ContainerDied","Data":"9a15994acf814d96da9fbcb84987a2edeb10af89e1b301670246344f9659e481"} Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.323300 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.369278 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7372c865-b00f-4772-89c8-3e5a8a662eb6-crc-storage\") pod \"7372c865-b00f-4772-89c8-3e5a8a662eb6\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.369361 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w27vt\" (UniqueName: \"kubernetes.io/projected/7372c865-b00f-4772-89c8-3e5a8a662eb6-kube-api-access-w27vt\") pod \"7372c865-b00f-4772-89c8-3e5a8a662eb6\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.369496 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7372c865-b00f-4772-89c8-3e5a8a662eb6-node-mnt\") pod \"7372c865-b00f-4772-89c8-3e5a8a662eb6\" (UID: \"7372c865-b00f-4772-89c8-3e5a8a662eb6\") " Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.369661 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7372c865-b00f-4772-89c8-3e5a8a662eb6-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "7372c865-b00f-4772-89c8-3e5a8a662eb6" (UID: "7372c865-b00f-4772-89c8-3e5a8a662eb6"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.369892 4944 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7372c865-b00f-4772-89c8-3e5a8a662eb6-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.376145 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7372c865-b00f-4772-89c8-3e5a8a662eb6-kube-api-access-w27vt" (OuterVolumeSpecName: "kube-api-access-w27vt") pod "7372c865-b00f-4772-89c8-3e5a8a662eb6" (UID: "7372c865-b00f-4772-89c8-3e5a8a662eb6"). InnerVolumeSpecName "kube-api-access-w27vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.387721 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7372c865-b00f-4772-89c8-3e5a8a662eb6-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "7372c865-b00f-4772-89c8-3e5a8a662eb6" (UID: "7372c865-b00f-4772-89c8-3e5a8a662eb6"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.471152 4944 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7372c865-b00f-4772-89c8-3e5a8a662eb6-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.471212 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w27vt\" (UniqueName: \"kubernetes.io/projected/7372c865-b00f-4772-89c8-3e5a8a662eb6-kube-api-access-w27vt\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.898736 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7q5tz"] Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.899139 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovn-controller" containerID="cri-o://c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152" gracePeriod=30 Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.899186 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="northd" containerID="cri-o://beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597" gracePeriod=30 Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.899256 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="kube-rbac-proxy-node" containerID="cri-o://304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f" gracePeriod=30 Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.899257 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovn-acl-logging" containerID="cri-o://3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43" gracePeriod=30 Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.899255 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="nbdb" containerID="cri-o://4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98" gracePeriod=30 Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.899316 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c" gracePeriod=30 Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.901629 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="sbdb" containerID="cri-o://2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1" gracePeriod=30 Nov 24 09:03:38 crc kubenswrapper[4944]: I1124 09:03:38.942664 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" containerID="cri-o://5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb" gracePeriod=30 Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.116383 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-462nz" event={"ID":"7372c865-b00f-4772-89c8-3e5a8a662eb6","Type":"ContainerDied","Data":"159b728deca143210b65f13ed48499156dd3307b636ea19873eedb3c5e7f6c27"} Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.116432 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="159b728deca143210b65f13ed48499156dd3307b636ea19873eedb3c5e7f6c27" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.116499 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-462nz" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.121352 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dg4dk_5ff38b2b-032d-47df-8836-105e0cfae835/kube-multus/2.log" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.121740 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dg4dk_5ff38b2b-032d-47df-8836-105e0cfae835/kube-multus/1.log" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.121773 4944 generic.go:334] "Generic (PLEG): container finished" podID="5ff38b2b-032d-47df-8836-105e0cfae835" containerID="010a18b9e549761ace384c33053c30fa54319e5b84f6acebd70957282468e242" exitCode=2 Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.121814 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dg4dk" event={"ID":"5ff38b2b-032d-47df-8836-105e0cfae835","Type":"ContainerDied","Data":"010a18b9e549761ace384c33053c30fa54319e5b84f6acebd70957282468e242"} Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.121842 4944 scope.go:117] "RemoveContainer" containerID="df7f5a8edddf8468d2a9558ded9fddc495508165211a6199d02e93dd26f64d0f" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.122215 4944 scope.go:117] "RemoveContainer" containerID="010a18b9e549761ace384c33053c30fa54319e5b84f6acebd70957282468e242" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.122477 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-dg4dk_openshift-multus(5ff38b2b-032d-47df-8836-105e0cfae835)\"" pod="openshift-multus/multus-dg4dk" podUID="5ff38b2b-032d-47df-8836-105e0cfae835" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.126390 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovnkube-controller/3.log" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.128731 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovn-acl-logging/0.log" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.129413 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovn-controller/0.log" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.129692 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb" exitCode=0 Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.129715 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c" exitCode=0 Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.129725 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f" exitCode=0 Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.129735 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43" exitCode=143 Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.129744 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152" exitCode=143 Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.129760 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb"} Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.129780 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c"} Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.129790 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f"} Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.129801 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43"} Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.129810 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152"} Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.163744 4944 scope.go:117] "RemoveContainer" containerID="e5ff1d37abc4f85624aa3e8a23d2a0fcd4225f42f7b627296fb40d8bd45675c3" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.178665 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovn-acl-logging/0.log" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.179238 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovn-controller/0.log" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.179630 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.227519 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-j8q4s"] Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.227749 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="nbdb" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.227765 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="nbdb" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.227775 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.227784 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.227793 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovn-acl-logging" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.227802 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovn-acl-logging" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.227811 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="northd" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.227820 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="northd" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.227832 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.227839 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.227849 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovn-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.227856 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovn-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.227865 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230218 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.230270 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="sbdb" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230282 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="sbdb" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.230311 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="kube-rbac-proxy-node" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230321 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="kube-rbac-proxy-node" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.230340 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7372c865-b00f-4772-89c8-3e5a8a662eb6" containerName="storage" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230348 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7372c865-b00f-4772-89c8-3e5a8a662eb6" containerName="storage" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.230357 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230365 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.230375 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="kubecfg-setup" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230384 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="kubecfg-setup" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230554 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230566 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230574 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="kube-rbac-proxy-node" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230584 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230593 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230601 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="7372c865-b00f-4772-89c8-3e5a8a662eb6" containerName="storage" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230607 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovn-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230614 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="sbdb" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230621 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230627 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230634 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="nbdb" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230642 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovn-acl-logging" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230651 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="northd" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.230736 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230744 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: E1124 09:03:39.230758 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.230764 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerName="ovnkube-controller" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.232359 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.279490 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-netns\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.279756 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-config\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.279599 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.279839 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-kubelet\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.279903 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-script-lib\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.279931 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-slash\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.279954 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-ovn-kubernetes\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.279984 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn57m\" (UniqueName: \"kubernetes.io/projected/aeb94161-d41a-420b-b5ed-9b65abbe91a3-kube-api-access-bn57m\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280004 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-systemd-units\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280005 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-slash" (OuterVolumeSpecName: "host-slash") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280023 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-etc-openvswitch\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280065 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-node-log\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280091 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovn-node-metrics-cert\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280111 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-systemd\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280135 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-netd\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280158 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-bin\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280188 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-env-overrides\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280207 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-ovn\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280229 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-openvswitch\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280246 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-var-lib-openvswitch\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280269 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280291 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-log-socket\") pod \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\" (UID: \"aeb94161-d41a-420b-b5ed-9b65abbe91a3\") " Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280365 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-kubelet\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280396 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-var-lib-openvswitch\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280718 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-log-socket\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280744 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-ovnkube-config\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280768 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-systemd-units\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280788 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-slash\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280811 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-etc-openvswitch\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280834 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-run-ovn\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280859 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-ovn-node-metrics-cert\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280879 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-cni-bin\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280909 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280243 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280342 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280375 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280395 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280417 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280442 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280460 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280477 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280605 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-node-log" (OuterVolumeSpecName: "node-log") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280876 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280900 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.280943 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281002 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281030 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-log-socket" (OuterVolumeSpecName: "log-socket") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281127 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-run-systemd\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281222 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjj2l\" (UniqueName: \"kubernetes.io/projected/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-kube-api-access-jjj2l\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281265 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-run-netns\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281286 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-env-overrides\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281307 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-run-openvswitch\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281328 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-cni-netd\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281353 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-node-log\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281388 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-run-ovn-kubernetes\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281409 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-ovnkube-script-lib\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281455 4944 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281469 4944 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281482 4944 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281493 4944 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281504 4944 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281515 4944 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281526 4944 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281538 4944 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281548 4944 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281559 4944 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281574 4944 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281584 4944 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aeb94161-d41a-420b-b5ed-9b65abbe91a3-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281594 4944 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281604 4944 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281615 4944 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.281627 4944 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.282302 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.284766 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.284818 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb94161-d41a-420b-b5ed-9b65abbe91a3-kube-api-access-bn57m" (OuterVolumeSpecName: "kube-api-access-bn57m") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "kube-api-access-bn57m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.292378 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "aeb94161-d41a-420b-b5ed-9b65abbe91a3" (UID: "aeb94161-d41a-420b-b5ed-9b65abbe91a3"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.382933 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.382990 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-run-systemd\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383073 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjj2l\" (UniqueName: \"kubernetes.io/projected/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-kube-api-access-jjj2l\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383065 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383143 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-run-systemd\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383104 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-run-netns\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383204 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-run-netns\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383231 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-env-overrides\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383266 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-run-openvswitch\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383280 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-cni-netd\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383308 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-node-log\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383380 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-run-ovn-kubernetes\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383406 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-ovnkube-script-lib\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383436 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-kubelet\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383459 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-var-lib-openvswitch\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383527 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-log-socket\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383560 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-ovnkube-config\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383588 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-slash\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383607 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-systemd-units\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383625 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-etc-openvswitch\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383687 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-run-ovn\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383721 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-ovn-node-metrics-cert\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383743 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-cni-bin\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383816 4944 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383833 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn57m\" (UniqueName: \"kubernetes.io/projected/aeb94161-d41a-420b-b5ed-9b65abbe91a3-kube-api-access-bn57m\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383849 4944 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aeb94161-d41a-420b-b5ed-9b65abbe91a3-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383860 4944 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aeb94161-d41a-420b-b5ed-9b65abbe91a3-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383928 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-env-overrides\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383988 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-node-log\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.383989 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-cni-bin\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384031 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-run-ovn-kubernetes\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384120 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-systemd-units\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384160 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-log-socket\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384480 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-etc-openvswitch\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384501 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-ovnkube-script-lib\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384540 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-kubelet\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384547 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-run-ovn\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384579 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-var-lib-openvswitch\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384605 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-run-openvswitch\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384631 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-cni-netd\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384662 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-host-slash\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.384743 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-ovnkube-config\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.387732 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-ovn-node-metrics-cert\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.397887 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjj2l\" (UniqueName: \"kubernetes.io/projected/f3b496ee-a7ed-4cad-98e0-ffeb14afaf50-kube-api-access-jjj2l\") pod \"ovnkube-node-j8q4s\" (UID: \"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50\") " pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: I1124 09:03:39.544649 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:39 crc kubenswrapper[4944]: W1124 09:03:39.561447 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3b496ee_a7ed_4cad_98e0_ffeb14afaf50.slice/crio-163bd19aac649477a551c92d4d1178af9e80e8319abddc6cb0cf55e79ae261c6 WatchSource:0}: Error finding container 163bd19aac649477a551c92d4d1178af9e80e8319abddc6cb0cf55e79ae261c6: Status 404 returned error can't find the container with id 163bd19aac649477a551c92d4d1178af9e80e8319abddc6cb0cf55e79ae261c6 Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.140480 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovn-acl-logging/0.log" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.141471 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7q5tz_aeb94161-d41a-420b-b5ed-9b65abbe91a3/ovn-controller/0.log" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.146376 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1" exitCode=0 Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.146424 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98" exitCode=0 Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.146432 4944 generic.go:334] "Generic (PLEG): container finished" podID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" containerID="beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597" exitCode=0 Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.146516 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1"} Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.146544 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98"} Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.146564 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597"} Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.146575 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" event={"ID":"aeb94161-d41a-420b-b5ed-9b65abbe91a3","Type":"ContainerDied","Data":"28466347089e5136be573f05db427e14a4068e3ceb329c7f08f3dc78984701a7"} Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.146593 4944 scope.go:117] "RemoveContainer" containerID="5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.146816 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7q5tz" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.148745 4944 generic.go:334] "Generic (PLEG): container finished" podID="f3b496ee-a7ed-4cad-98e0-ffeb14afaf50" containerID="ec6c6652b405cee44267518fc6da08bc69430ed75bc4aaada3700bb3eadf2fe4" exitCode=0 Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.148859 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" event={"ID":"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50","Type":"ContainerDied","Data":"ec6c6652b405cee44267518fc6da08bc69430ed75bc4aaada3700bb3eadf2fe4"} Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.148892 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" event={"ID":"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50","Type":"ContainerStarted","Data":"163bd19aac649477a551c92d4d1178af9e80e8319abddc6cb0cf55e79ae261c6"} Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.152575 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dg4dk_5ff38b2b-032d-47df-8836-105e0cfae835/kube-multus/2.log" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.178305 4944 scope.go:117] "RemoveContainer" containerID="2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.199202 4944 scope.go:117] "RemoveContainer" containerID="4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.219281 4944 scope.go:117] "RemoveContainer" containerID="beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.223209 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7q5tz"] Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.225028 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7q5tz"] Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.244322 4944 scope.go:117] "RemoveContainer" containerID="1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.256891 4944 scope.go:117] "RemoveContainer" containerID="304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.269923 4944 scope.go:117] "RemoveContainer" containerID="3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.284297 4944 scope.go:117] "RemoveContainer" containerID="c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.292129 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeb94161-d41a-420b-b5ed-9b65abbe91a3" path="/var/lib/kubelet/pods/aeb94161-d41a-420b-b5ed-9b65abbe91a3/volumes" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.317646 4944 scope.go:117] "RemoveContainer" containerID="444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.342277 4944 scope.go:117] "RemoveContainer" containerID="5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb" Nov 24 09:03:40 crc kubenswrapper[4944]: E1124 09:03:40.342635 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb\": container with ID starting with 5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb not found: ID does not exist" containerID="5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.342664 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb"} err="failed to get container status \"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb\": rpc error: code = NotFound desc = could not find container \"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb\": container with ID starting with 5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.342683 4944 scope.go:117] "RemoveContainer" containerID="2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1" Nov 24 09:03:40 crc kubenswrapper[4944]: E1124 09:03:40.342868 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\": container with ID starting with 2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1 not found: ID does not exist" containerID="2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.342886 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1"} err="failed to get container status \"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\": rpc error: code = NotFound desc = could not find container \"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\": container with ID starting with 2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.342899 4944 scope.go:117] "RemoveContainer" containerID="4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98" Nov 24 09:03:40 crc kubenswrapper[4944]: E1124 09:03:40.343077 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\": container with ID starting with 4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98 not found: ID does not exist" containerID="4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.343094 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98"} err="failed to get container status \"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\": rpc error: code = NotFound desc = could not find container \"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\": container with ID starting with 4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.343110 4944 scope.go:117] "RemoveContainer" containerID="beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597" Nov 24 09:03:40 crc kubenswrapper[4944]: E1124 09:03:40.343275 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\": container with ID starting with beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597 not found: ID does not exist" containerID="beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.343292 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597"} err="failed to get container status \"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\": rpc error: code = NotFound desc = could not find container \"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\": container with ID starting with beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.343304 4944 scope.go:117] "RemoveContainer" containerID="1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c" Nov 24 09:03:40 crc kubenswrapper[4944]: E1124 09:03:40.343488 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\": container with ID starting with 1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c not found: ID does not exist" containerID="1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.343559 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c"} err="failed to get container status \"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\": rpc error: code = NotFound desc = could not find container \"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\": container with ID starting with 1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.343582 4944 scope.go:117] "RemoveContainer" containerID="304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f" Nov 24 09:03:40 crc kubenswrapper[4944]: E1124 09:03:40.343746 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\": container with ID starting with 304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f not found: ID does not exist" containerID="304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.343761 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f"} err="failed to get container status \"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\": rpc error: code = NotFound desc = could not find container \"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\": container with ID starting with 304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.343773 4944 scope.go:117] "RemoveContainer" containerID="3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43" Nov 24 09:03:40 crc kubenswrapper[4944]: E1124 09:03:40.343959 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\": container with ID starting with 3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43 not found: ID does not exist" containerID="3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.343975 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43"} err="failed to get container status \"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\": rpc error: code = NotFound desc = could not find container \"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\": container with ID starting with 3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.343988 4944 scope.go:117] "RemoveContainer" containerID="c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152" Nov 24 09:03:40 crc kubenswrapper[4944]: E1124 09:03:40.344250 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\": container with ID starting with c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152 not found: ID does not exist" containerID="c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.344294 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152"} err="failed to get container status \"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\": rpc error: code = NotFound desc = could not find container \"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\": container with ID starting with c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.344310 4944 scope.go:117] "RemoveContainer" containerID="444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4" Nov 24 09:03:40 crc kubenswrapper[4944]: E1124 09:03:40.344607 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\": container with ID starting with 444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4 not found: ID does not exist" containerID="444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.344631 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4"} err="failed to get container status \"444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\": rpc error: code = NotFound desc = could not find container \"444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\": container with ID starting with 444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.344647 4944 scope.go:117] "RemoveContainer" containerID="5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.345010 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb"} err="failed to get container status \"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb\": rpc error: code = NotFound desc = could not find container \"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb\": container with ID starting with 5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.345075 4944 scope.go:117] "RemoveContainer" containerID="2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.345475 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1"} err="failed to get container status \"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\": rpc error: code = NotFound desc = could not find container \"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\": container with ID starting with 2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.345493 4944 scope.go:117] "RemoveContainer" containerID="4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.345719 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98"} err="failed to get container status \"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\": rpc error: code = NotFound desc = could not find container \"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\": container with ID starting with 4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.345734 4944 scope.go:117] "RemoveContainer" containerID="beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.345956 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597"} err="failed to get container status \"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\": rpc error: code = NotFound desc = could not find container \"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\": container with ID starting with beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.345978 4944 scope.go:117] "RemoveContainer" containerID="1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.346351 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c"} err="failed to get container status \"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\": rpc error: code = NotFound desc = could not find container \"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\": container with ID starting with 1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.346374 4944 scope.go:117] "RemoveContainer" containerID="304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.346608 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f"} err="failed to get container status \"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\": rpc error: code = NotFound desc = could not find container \"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\": container with ID starting with 304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.346635 4944 scope.go:117] "RemoveContainer" containerID="3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.346841 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43"} err="failed to get container status \"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\": rpc error: code = NotFound desc = could not find container \"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\": container with ID starting with 3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.346858 4944 scope.go:117] "RemoveContainer" containerID="c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.347086 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152"} err="failed to get container status \"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\": rpc error: code = NotFound desc = could not find container \"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\": container with ID starting with c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.347105 4944 scope.go:117] "RemoveContainer" containerID="444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.347324 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4"} err="failed to get container status \"444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\": rpc error: code = NotFound desc = could not find container \"444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\": container with ID starting with 444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.347340 4944 scope.go:117] "RemoveContainer" containerID="5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.347576 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb"} err="failed to get container status \"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb\": rpc error: code = NotFound desc = could not find container \"5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb\": container with ID starting with 5cccfb1025092f24ef8cba4e6e967eab4c28fc9c7c67017c5a42b44a879e6feb not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.347592 4944 scope.go:117] "RemoveContainer" containerID="2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.347812 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1"} err="failed to get container status \"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\": rpc error: code = NotFound desc = could not find container \"2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1\": container with ID starting with 2259b452f7255ab6e22fc7287a9c81fb5263c3e64986ebfa8174cbc00e2a0ff1 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.347838 4944 scope.go:117] "RemoveContainer" containerID="4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.348200 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98"} err="failed to get container status \"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\": rpc error: code = NotFound desc = could not find container \"4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98\": container with ID starting with 4847067b5b113adceaa338b513384b5c82242f30abb4628b41fb6c9879eb6a98 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.348224 4944 scope.go:117] "RemoveContainer" containerID="beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.351411 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597"} err="failed to get container status \"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\": rpc error: code = NotFound desc = could not find container \"beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597\": container with ID starting with beb9d85335eb615e46b3367031b1bf48dcb73b83deaa1b395d12977c7b4a1597 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.351432 4944 scope.go:117] "RemoveContainer" containerID="1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.351745 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c"} err="failed to get container status \"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\": rpc error: code = NotFound desc = could not find container \"1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c\": container with ID starting with 1173f0323cbc4a2751a796871a8800ae82500fb82b045a2717c1813d2695d20c not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.351776 4944 scope.go:117] "RemoveContainer" containerID="304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.352611 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f"} err="failed to get container status \"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\": rpc error: code = NotFound desc = could not find container \"304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f\": container with ID starting with 304729a141bf1ec6737c0ca17da501637e101ab7abde266dcac03fcffbf8528f not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.352632 4944 scope.go:117] "RemoveContainer" containerID="3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.352971 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43"} err="failed to get container status \"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\": rpc error: code = NotFound desc = could not find container \"3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43\": container with ID starting with 3f9c9afd852944a0b8e02ebb87774bce72ede8e0c18016ad4d9715824d9dbb43 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.353011 4944 scope.go:117] "RemoveContainer" containerID="c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.353317 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152"} err="failed to get container status \"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\": rpc error: code = NotFound desc = could not find container \"c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152\": container with ID starting with c6d1a55ad20939fbaa897d18f1135cc93784209ffdacf191ea28507f343eb152 not found: ID does not exist" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.353339 4944 scope.go:117] "RemoveContainer" containerID="444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4" Nov 24 09:03:40 crc kubenswrapper[4944]: I1124 09:03:40.353712 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4"} err="failed to get container status \"444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\": rpc error: code = NotFound desc = could not find container \"444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4\": container with ID starting with 444d7ec4b369879bf46f6f1fb33218e701d11615be57792ce6cf3e20199dfce4 not found: ID does not exist" Nov 24 09:03:41 crc kubenswrapper[4944]: I1124 09:03:41.167244 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" event={"ID":"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50","Type":"ContainerStarted","Data":"f14e8983e53b865e27004544f819575f0c41ac5a650d51b7d3434cdb2016b0a8"} Nov 24 09:03:41 crc kubenswrapper[4944]: I1124 09:03:41.167774 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" event={"ID":"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50","Type":"ContainerStarted","Data":"da3048d99fe3623bd713a50fcf49e52ad8a231b1e5d471154c8359861d19fe7f"} Nov 24 09:03:41 crc kubenswrapper[4944]: I1124 09:03:41.167792 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" event={"ID":"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50","Type":"ContainerStarted","Data":"cd44d7b62e625b133ee9222f08760a8b995bd9cc2651141ef6e36b3891c7854a"} Nov 24 09:03:41 crc kubenswrapper[4944]: I1124 09:03:41.167837 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" event={"ID":"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50","Type":"ContainerStarted","Data":"50dd38568ad7fee39c03145f6fe90d9ef7696723f050a0cfd9e4a032871f1589"} Nov 24 09:03:41 crc kubenswrapper[4944]: I1124 09:03:41.167851 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" event={"ID":"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50","Type":"ContainerStarted","Data":"01c6d4b0d01a38b44a8e950ef022429ba6a95017e8be18717b60e363530d4a47"} Nov 24 09:03:41 crc kubenswrapper[4944]: I1124 09:03:41.167863 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" event={"ID":"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50","Type":"ContainerStarted","Data":"27714110a5a6dc8a8378a5ab86a2c8b6b7f272a2a3f4019f111773864bdce2d7"} Nov 24 09:03:43 crc kubenswrapper[4944]: I1124 09:03:43.184544 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" event={"ID":"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50","Type":"ContainerStarted","Data":"53d21b268089a30884d588f7e31b4a49e33aab37dedf91127e22bec873ba0f4d"} Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.279472 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49"] Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.281165 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.283491 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.454136 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.454254 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9528\" (UniqueName: \"kubernetes.io/projected/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-kube-api-access-w9528\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.454312 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.555562 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9528\" (UniqueName: \"kubernetes.io/projected/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-kube-api-access-w9528\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.555630 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.555658 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.556170 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.556424 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.572204 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9528\" (UniqueName: \"kubernetes.io/projected/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-kube-api-access-w9528\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.599375 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: E1124 09:03:45.621488 4944 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(c04d9f8476a7342d0ae9dd55e5490d95856b0c789ca62d7523f1115c6dbccfcf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 09:03:45 crc kubenswrapper[4944]: E1124 09:03:45.621575 4944 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(c04d9f8476a7342d0ae9dd55e5490d95856b0c789ca62d7523f1115c6dbccfcf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: E1124 09:03:45.621603 4944 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(c04d9f8476a7342d0ae9dd55e5490d95856b0c789ca62d7523f1115c6dbccfcf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:45 crc kubenswrapper[4944]: E1124 09:03:45.621662 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace(66b08e16-cf20-4ebb-a6b0-e33bd5947b75)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace(66b08e16-cf20-4ebb-a6b0-e33bd5947b75)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(c04d9f8476a7342d0ae9dd55e5490d95856b0c789ca62d7523f1115c6dbccfcf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" podUID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" Nov 24 09:03:45 crc kubenswrapper[4944]: I1124 09:03:45.961814 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49"] Nov 24 09:03:46 crc kubenswrapper[4944]: I1124 09:03:46.205287 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:46 crc kubenswrapper[4944]: I1124 09:03:46.205291 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" event={"ID":"f3b496ee-a7ed-4cad-98e0-ffeb14afaf50","Type":"ContainerStarted","Data":"b7b0d9afef844a2a94c74dd3f09915234578a52ddaa3e6fc7904b4ec9ba71fcd"} Nov 24 09:03:46 crc kubenswrapper[4944]: I1124 09:03:46.205887 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:46 crc kubenswrapper[4944]: I1124 09:03:46.205936 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:46 crc kubenswrapper[4944]: I1124 09:03:46.206204 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:46 crc kubenswrapper[4944]: I1124 09:03:46.234395 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:46 crc kubenswrapper[4944]: E1124 09:03:46.236980 4944 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(9d27c67f0e8c7b67ad3ebe90cf38d341c6c54ebc14e7d02c95c8092dac8a2760): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 09:03:46 crc kubenswrapper[4944]: E1124 09:03:46.237023 4944 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(9d27c67f0e8c7b67ad3ebe90cf38d341c6c54ebc14e7d02c95c8092dac8a2760): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:46 crc kubenswrapper[4944]: E1124 09:03:46.237060 4944 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(9d27c67f0e8c7b67ad3ebe90cf38d341c6c54ebc14e7d02c95c8092dac8a2760): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:46 crc kubenswrapper[4944]: E1124 09:03:46.237101 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace(66b08e16-cf20-4ebb-a6b0-e33bd5947b75)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace(66b08e16-cf20-4ebb-a6b0-e33bd5947b75)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(9d27c67f0e8c7b67ad3ebe90cf38d341c6c54ebc14e7d02c95c8092dac8a2760): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" podUID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" Nov 24 09:03:46 crc kubenswrapper[4944]: I1124 09:03:46.239210 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" podStartSLOduration=7.239196094 podStartE2EDuration="7.239196094s" podCreationTimestamp="2025-11-24 09:03:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:03:46.23750783 +0000 UTC m=+686.771948292" watchObservedRunningTime="2025-11-24 09:03:46.239196094 +0000 UTC m=+686.773636556" Nov 24 09:03:47 crc kubenswrapper[4944]: I1124 09:03:47.210293 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:47 crc kubenswrapper[4944]: I1124 09:03:47.247741 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:03:53 crc kubenswrapper[4944]: I1124 09:03:53.276914 4944 scope.go:117] "RemoveContainer" containerID="010a18b9e549761ace384c33053c30fa54319e5b84f6acebd70957282468e242" Nov 24 09:03:53 crc kubenswrapper[4944]: E1124 09:03:53.277724 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-dg4dk_openshift-multus(5ff38b2b-032d-47df-8836-105e0cfae835)\"" pod="openshift-multus/multus-dg4dk" podUID="5ff38b2b-032d-47df-8836-105e0cfae835" Nov 24 09:03:53 crc kubenswrapper[4944]: I1124 09:03:53.548213 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:03:53 crc kubenswrapper[4944]: I1124 09:03:53.548277 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:03:59 crc kubenswrapper[4944]: I1124 09:03:59.276121 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:59 crc kubenswrapper[4944]: I1124 09:03:59.277626 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:59 crc kubenswrapper[4944]: E1124 09:03:59.300040 4944 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(851c1ead0729411f0b18ae8b1ffd6bfdd6ef008adfd87bf0ba0d52cf892c1d83): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 09:03:59 crc kubenswrapper[4944]: E1124 09:03:59.300118 4944 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(851c1ead0729411f0b18ae8b1ffd6bfdd6ef008adfd87bf0ba0d52cf892c1d83): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:59 crc kubenswrapper[4944]: E1124 09:03:59.300144 4944 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(851c1ead0729411f0b18ae8b1ffd6bfdd6ef008adfd87bf0ba0d52cf892c1d83): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:03:59 crc kubenswrapper[4944]: E1124 09:03:59.300184 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace(66b08e16-cf20-4ebb-a6b0-e33bd5947b75)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace(66b08e16-cf20-4ebb-a6b0-e33bd5947b75)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_openshift-marketplace_66b08e16-cf20-4ebb-a6b0-e33bd5947b75_0(851c1ead0729411f0b18ae8b1ffd6bfdd6ef008adfd87bf0ba0d52cf892c1d83): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" podUID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" Nov 24 09:04:07 crc kubenswrapper[4944]: I1124 09:04:07.277200 4944 scope.go:117] "RemoveContainer" containerID="010a18b9e549761ace384c33053c30fa54319e5b84f6acebd70957282468e242" Nov 24 09:04:08 crc kubenswrapper[4944]: I1124 09:04:08.308622 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dg4dk_5ff38b2b-032d-47df-8836-105e0cfae835/kube-multus/2.log" Nov 24 09:04:08 crc kubenswrapper[4944]: I1124 09:04:08.308993 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dg4dk" event={"ID":"5ff38b2b-032d-47df-8836-105e0cfae835","Type":"ContainerStarted","Data":"b8735c3fb7f418bfca3d2b47dbad45d27223a19e25a01a8ff24c8e9b5ea1cd03"} Nov 24 09:04:09 crc kubenswrapper[4944]: I1124 09:04:09.562975 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-j8q4s" Nov 24 09:04:14 crc kubenswrapper[4944]: I1124 09:04:14.276521 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:04:14 crc kubenswrapper[4944]: I1124 09:04:14.276955 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:04:14 crc kubenswrapper[4944]: I1124 09:04:14.452163 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49"] Nov 24 09:04:15 crc kubenswrapper[4944]: I1124 09:04:15.343142 4944 generic.go:334] "Generic (PLEG): container finished" podID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" containerID="af87f5eab1e9b421e06c6b106442c22c82248ee04349982b05500b6821026e75" exitCode=0 Nov 24 09:04:15 crc kubenswrapper[4944]: I1124 09:04:15.343244 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" event={"ID":"66b08e16-cf20-4ebb-a6b0-e33bd5947b75","Type":"ContainerDied","Data":"af87f5eab1e9b421e06c6b106442c22c82248ee04349982b05500b6821026e75"} Nov 24 09:04:15 crc kubenswrapper[4944]: I1124 09:04:15.344145 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" event={"ID":"66b08e16-cf20-4ebb-a6b0-e33bd5947b75","Type":"ContainerStarted","Data":"6b2a0fdd20b2cdb67c1d1c91d42e55e5ce28688508b593997dd8566e4a138d36"} Nov 24 09:04:17 crc kubenswrapper[4944]: I1124 09:04:17.363368 4944 generic.go:334] "Generic (PLEG): container finished" podID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" containerID="0c0fc4f13a0c1fee0a28268df0fb26dfb7590173a21b1c4ca8a0d4dcfb3dc870" exitCode=0 Nov 24 09:04:17 crc kubenswrapper[4944]: I1124 09:04:17.363502 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" event={"ID":"66b08e16-cf20-4ebb-a6b0-e33bd5947b75","Type":"ContainerDied","Data":"0c0fc4f13a0c1fee0a28268df0fb26dfb7590173a21b1c4ca8a0d4dcfb3dc870"} Nov 24 09:04:18 crc kubenswrapper[4944]: I1124 09:04:18.371352 4944 generic.go:334] "Generic (PLEG): container finished" podID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" containerID="549ec6b2ada5345a8560306120abee69e97d1711cbd1b98036cafb96fab71273" exitCode=0 Nov 24 09:04:18 crc kubenswrapper[4944]: I1124 09:04:18.371431 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" event={"ID":"66b08e16-cf20-4ebb-a6b0-e33bd5947b75","Type":"ContainerDied","Data":"549ec6b2ada5345a8560306120abee69e97d1711cbd1b98036cafb96fab71273"} Nov 24 09:04:19 crc kubenswrapper[4944]: I1124 09:04:19.581703 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:04:19 crc kubenswrapper[4944]: I1124 09:04:19.764373 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9528\" (UniqueName: \"kubernetes.io/projected/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-kube-api-access-w9528\") pod \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " Nov 24 09:04:19 crc kubenswrapper[4944]: I1124 09:04:19.764509 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-bundle\") pod \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " Nov 24 09:04:19 crc kubenswrapper[4944]: I1124 09:04:19.764549 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-util\") pod \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\" (UID: \"66b08e16-cf20-4ebb-a6b0-e33bd5947b75\") " Nov 24 09:04:19 crc kubenswrapper[4944]: I1124 09:04:19.765310 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-bundle" (OuterVolumeSpecName: "bundle") pod "66b08e16-cf20-4ebb-a6b0-e33bd5947b75" (UID: "66b08e16-cf20-4ebb-a6b0-e33bd5947b75"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:04:19 crc kubenswrapper[4944]: I1124 09:04:19.772171 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-kube-api-access-w9528" (OuterVolumeSpecName: "kube-api-access-w9528") pod "66b08e16-cf20-4ebb-a6b0-e33bd5947b75" (UID: "66b08e16-cf20-4ebb-a6b0-e33bd5947b75"). InnerVolumeSpecName "kube-api-access-w9528". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:04:19 crc kubenswrapper[4944]: I1124 09:04:19.866317 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9528\" (UniqueName: \"kubernetes.io/projected/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-kube-api-access-w9528\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:19 crc kubenswrapper[4944]: I1124 09:04:19.866359 4944 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:20 crc kubenswrapper[4944]: I1124 09:04:20.033386 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-util" (OuterVolumeSpecName: "util") pod "66b08e16-cf20-4ebb-a6b0-e33bd5947b75" (UID: "66b08e16-cf20-4ebb-a6b0-e33bd5947b75"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:04:20 crc kubenswrapper[4944]: I1124 09:04:20.068832 4944 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66b08e16-cf20-4ebb-a6b0-e33bd5947b75-util\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:20 crc kubenswrapper[4944]: I1124 09:04:20.388347 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" event={"ID":"66b08e16-cf20-4ebb-a6b0-e33bd5947b75","Type":"ContainerDied","Data":"6b2a0fdd20b2cdb67c1d1c91d42e55e5ce28688508b593997dd8566e4a138d36"} Nov 24 09:04:20 crc kubenswrapper[4944]: I1124 09:04:20.388414 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b2a0fdd20b2cdb67c1d1c91d42e55e5ce28688508b593997dd8566e4a138d36" Nov 24 09:04:20 crc kubenswrapper[4944]: I1124 09:04:20.388420 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49" Nov 24 09:04:23 crc kubenswrapper[4944]: I1124 09:04:23.548390 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:04:23 crc kubenswrapper[4944]: I1124 09:04:23.548700 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.011819 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-hzrsn"] Nov 24 09:04:27 crc kubenswrapper[4944]: E1124 09:04:27.012240 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" containerName="util" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.012251 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" containerName="util" Nov 24 09:04:27 crc kubenswrapper[4944]: E1124 09:04:27.012261 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" containerName="extract" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.012267 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" containerName="extract" Nov 24 09:04:27 crc kubenswrapper[4944]: E1124 09:04:27.012285 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" containerName="pull" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.012291 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" containerName="pull" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.012372 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b08e16-cf20-4ebb-a6b0-e33bd5947b75" containerName="extract" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.012687 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-hzrsn" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.014844 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.015839 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.016253 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-s6bzl" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.022622 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-hzrsn"] Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.152939 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxcsb\" (UniqueName: \"kubernetes.io/projected/3791dd89-b8cd-4a94-9542-3929fe9a9872-kube-api-access-vxcsb\") pod \"nmstate-operator-557fdffb88-hzrsn\" (UID: \"3791dd89-b8cd-4a94-9542-3929fe9a9872\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-hzrsn" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.254848 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxcsb\" (UniqueName: \"kubernetes.io/projected/3791dd89-b8cd-4a94-9542-3929fe9a9872-kube-api-access-vxcsb\") pod \"nmstate-operator-557fdffb88-hzrsn\" (UID: \"3791dd89-b8cd-4a94-9542-3929fe9a9872\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-hzrsn" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.274125 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxcsb\" (UniqueName: \"kubernetes.io/projected/3791dd89-b8cd-4a94-9542-3929fe9a9872-kube-api-access-vxcsb\") pod \"nmstate-operator-557fdffb88-hzrsn\" (UID: \"3791dd89-b8cd-4a94-9542-3929fe9a9872\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-hzrsn" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.329707 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-hzrsn" Nov 24 09:04:27 crc kubenswrapper[4944]: I1124 09:04:27.524131 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-hzrsn"] Nov 24 09:04:27 crc kubenswrapper[4944]: E1124 09:04:27.880596 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice/crio-6b2a0fdd20b2cdb67c1d1c91d42e55e5ce28688508b593997dd8566e4a138d36\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice\": RecentStats: unable to find data in memory cache]" Nov 24 09:04:28 crc kubenswrapper[4944]: I1124 09:04:28.430452 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-hzrsn" event={"ID":"3791dd89-b8cd-4a94-9542-3929fe9a9872","Type":"ContainerStarted","Data":"d74c10b51acbfec1abb8bc0002804aa843454548bc2df277d0d0f22f1cb89657"} Nov 24 09:04:30 crc kubenswrapper[4944]: I1124 09:04:30.442467 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-hzrsn" event={"ID":"3791dd89-b8cd-4a94-9542-3929fe9a9872","Type":"ContainerStarted","Data":"526f87dd593ef605cdd26631c78bbebf0cea1c5c5424a764c56afd16aaf2d1a0"} Nov 24 09:04:30 crc kubenswrapper[4944]: I1124 09:04:30.458596 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-hzrsn" podStartSLOduration=2.529839569 podStartE2EDuration="4.458580581s" podCreationTimestamp="2025-11-24 09:04:26 +0000 UTC" firstStartedPulling="2025-11-24 09:04:27.539522369 +0000 UTC m=+728.073962831" lastFinishedPulling="2025-11-24 09:04:29.468263381 +0000 UTC m=+730.002703843" observedRunningTime="2025-11-24 09:04:30.457274599 +0000 UTC m=+730.991715061" watchObservedRunningTime="2025-11-24 09:04:30.458580581 +0000 UTC m=+730.993021043" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.022686 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz"] Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.023770 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.026110 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-27kt4" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.028816 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb"] Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.029350 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.030787 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.039211 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz"] Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.044614 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb"] Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.058763 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-vjtqb"] Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.059405 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.144434 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z"] Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.145077 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.151721 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.152010 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-gfskf" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.152374 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.156087 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z"] Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.171032 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6a5e82c5-f9a6-45ea-b831-36c6573b974e-nmstate-lock\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.171094 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npwpl\" (UniqueName: \"kubernetes.io/projected/299a592e-b90c-436a-9aa4-c2b949cb0e85-kube-api-access-npwpl\") pod \"nmstate-metrics-5dcf9c57c5-pqwqz\" (UID: \"299a592e-b90c-436a-9aa4-c2b949cb0e85\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.171136 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6a5e82c5-f9a6-45ea-b831-36c6573b974e-ovs-socket\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.171159 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3f916409-7217-43dd-a0fe-859fce69591e-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-sxnnb\" (UID: \"3f916409-7217-43dd-a0fe-859fce69591e\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.171177 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9vgd\" (UniqueName: \"kubernetes.io/projected/6a5e82c5-f9a6-45ea-b831-36c6573b974e-kube-api-access-t9vgd\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.171191 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv9j8\" (UniqueName: \"kubernetes.io/projected/3f916409-7217-43dd-a0fe-859fce69591e-kube-api-access-gv9j8\") pod \"nmstate-webhook-6b89b748d8-sxnnb\" (UID: \"3f916409-7217-43dd-a0fe-859fce69591e\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.171207 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6a5e82c5-f9a6-45ea-b831-36c6573b974e-dbus-socket\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.272869 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk6m7\" (UniqueName: \"kubernetes.io/projected/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-kube-api-access-gk6m7\") pod \"nmstate-console-plugin-5874bd7bc5-74c8z\" (UID: \"1a7b278b-3ae9-4b09-b586-875bb2cac1d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.272931 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6a5e82c5-f9a6-45ea-b831-36c6573b974e-nmstate-lock\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.272960 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npwpl\" (UniqueName: \"kubernetes.io/projected/299a592e-b90c-436a-9aa4-c2b949cb0e85-kube-api-access-npwpl\") pod \"nmstate-metrics-5dcf9c57c5-pqwqz\" (UID: \"299a592e-b90c-436a-9aa4-c2b949cb0e85\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.273014 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6a5e82c5-f9a6-45ea-b831-36c6573b974e-ovs-socket\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.273037 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-74c8z\" (UID: \"1a7b278b-3ae9-4b09-b586-875bb2cac1d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.273087 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3f916409-7217-43dd-a0fe-859fce69591e-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-sxnnb\" (UID: \"3f916409-7217-43dd-a0fe-859fce69591e\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.273102 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6a5e82c5-f9a6-45ea-b831-36c6573b974e-nmstate-lock\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.273114 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-74c8z\" (UID: \"1a7b278b-3ae9-4b09-b586-875bb2cac1d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.273290 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9vgd\" (UniqueName: \"kubernetes.io/projected/6a5e82c5-f9a6-45ea-b831-36c6573b974e-kube-api-access-t9vgd\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.273174 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6a5e82c5-f9a6-45ea-b831-36c6573b974e-ovs-socket\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.273372 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv9j8\" (UniqueName: \"kubernetes.io/projected/3f916409-7217-43dd-a0fe-859fce69591e-kube-api-access-gv9j8\") pod \"nmstate-webhook-6b89b748d8-sxnnb\" (UID: \"3f916409-7217-43dd-a0fe-859fce69591e\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.273774 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6a5e82c5-f9a6-45ea-b831-36c6573b974e-dbus-socket\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.274060 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6a5e82c5-f9a6-45ea-b831-36c6573b974e-dbus-socket\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.288982 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3f916409-7217-43dd-a0fe-859fce69591e-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-sxnnb\" (UID: \"3f916409-7217-43dd-a0fe-859fce69591e\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.294863 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npwpl\" (UniqueName: \"kubernetes.io/projected/299a592e-b90c-436a-9aa4-c2b949cb0e85-kube-api-access-npwpl\") pod \"nmstate-metrics-5dcf9c57c5-pqwqz\" (UID: \"299a592e-b90c-436a-9aa4-c2b949cb0e85\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.297661 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9vgd\" (UniqueName: \"kubernetes.io/projected/6a5e82c5-f9a6-45ea-b831-36c6573b974e-kube-api-access-t9vgd\") pod \"nmstate-handler-vjtqb\" (UID: \"6a5e82c5-f9a6-45ea-b831-36c6573b974e\") " pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.299238 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv9j8\" (UniqueName: \"kubernetes.io/projected/3f916409-7217-43dd-a0fe-859fce69591e-kube-api-access-gv9j8\") pod \"nmstate-webhook-6b89b748d8-sxnnb\" (UID: \"3f916409-7217-43dd-a0fe-859fce69591e\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.337589 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-fdbf99894-8q8pr"] Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.338437 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.344715 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.357189 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.361404 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-fdbf99894-8q8pr"] Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.375399 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-74c8z\" (UID: \"1a7b278b-3ae9-4b09-b586-875bb2cac1d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.375492 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-trusted-ca-bundle\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.375530 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-74c8z\" (UID: \"1a7b278b-3ae9-4b09-b586-875bb2cac1d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.375557 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-oauth-serving-cert\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.375584 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-console-serving-cert\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.375626 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-console-config\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.375658 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-service-ca\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.375686 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk6m7\" (UniqueName: \"kubernetes.io/projected/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-kube-api-access-gk6m7\") pod \"nmstate-console-plugin-5874bd7bc5-74c8z\" (UID: \"1a7b278b-3ae9-4b09-b586-875bb2cac1d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.375709 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8p7m\" (UniqueName: \"kubernetes.io/projected/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-kube-api-access-q8p7m\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.375778 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-console-oauth-config\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.377361 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-74c8z\" (UID: \"1a7b278b-3ae9-4b09-b586-875bb2cac1d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:36 crc kubenswrapper[4944]: E1124 09:04:36.378485 4944 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 24 09:04:36 crc kubenswrapper[4944]: E1124 09:04:36.378538 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-plugin-serving-cert podName:1a7b278b-3ae9-4b09-b586-875bb2cac1d6 nodeName:}" failed. No retries permitted until 2025-11-24 09:04:36.878520388 +0000 UTC m=+737.412960930 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-plugin-serving-cert") pod "nmstate-console-plugin-5874bd7bc5-74c8z" (UID: "1a7b278b-3ae9-4b09-b586-875bb2cac1d6") : secret "plugin-serving-cert" not found Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.379129 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.396197 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk6m7\" (UniqueName: \"kubernetes.io/projected/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-kube-api-access-gk6m7\") pod \"nmstate-console-plugin-5874bd7bc5-74c8z\" (UID: \"1a7b278b-3ae9-4b09-b586-875bb2cac1d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.476452 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-service-ca\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.476495 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8p7m\" (UniqueName: \"kubernetes.io/projected/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-kube-api-access-q8p7m\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.476525 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-console-oauth-config\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.476562 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-trusted-ca-bundle\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.476591 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-oauth-serving-cert\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.476608 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-console-serving-cert\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.476631 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-console-config\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.478236 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-console-config\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.485843 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-service-ca\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.486654 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-oauth-serving-cert\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.488320 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-trusted-ca-bundle\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.489144 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-console-oauth-config\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.492072 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vjtqb" event={"ID":"6a5e82c5-f9a6-45ea-b831-36c6573b974e","Type":"ContainerStarted","Data":"ec080253f77ea478378cf2b023a93789157e4cd864c59c507acab42144f2334c"} Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.494510 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-console-serving-cert\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.506584 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8p7m\" (UniqueName: \"kubernetes.io/projected/e6b4d1cd-da30-49de-a548-5e57f4aaa7c7-kube-api-access-q8p7m\") pod \"console-fdbf99894-8q8pr\" (UID: \"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7\") " pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.593619 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb"] Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.635206 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz"] Nov 24 09:04:36 crc kubenswrapper[4944]: W1124 09:04:36.637457 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod299a592e_b90c_436a_9aa4_c2b949cb0e85.slice/crio-609b818d7be9d935f37746e381aa0211dbd6fafa8cc35ec1afd51e54122a1e13 WatchSource:0}: Error finding container 609b818d7be9d935f37746e381aa0211dbd6fafa8cc35ec1afd51e54122a1e13: Status 404 returned error can't find the container with id 609b818d7be9d935f37746e381aa0211dbd6fafa8cc35ec1afd51e54122a1e13 Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.705084 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.879666 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-74c8z\" (UID: \"1a7b278b-3ae9-4b09-b586-875bb2cac1d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.885605 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1a7b278b-3ae9-4b09-b586-875bb2cac1d6-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-74c8z\" (UID: \"1a7b278b-3ae9-4b09-b586-875bb2cac1d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:36 crc kubenswrapper[4944]: I1124 09:04:36.888430 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-fdbf99894-8q8pr"] Nov 24 09:04:37 crc kubenswrapper[4944]: I1124 09:04:37.064392 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" Nov 24 09:04:37 crc kubenswrapper[4944]: I1124 09:04:37.298518 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z"] Nov 24 09:04:37 crc kubenswrapper[4944]: I1124 09:04:37.500453 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" event={"ID":"3f916409-7217-43dd-a0fe-859fce69591e","Type":"ContainerStarted","Data":"ec7a10db6b9adb56ceb83577c78f55e8a85e47a379699da5da3f7208c48cad02"} Nov 24 09:04:37 crc kubenswrapper[4944]: I1124 09:04:37.501604 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" event={"ID":"1a7b278b-3ae9-4b09-b586-875bb2cac1d6","Type":"ContainerStarted","Data":"24e8aa5abc676b07a28c0723ab1f0d8f6d4701127cf9b091937d9e2502267257"} Nov 24 09:04:37 crc kubenswrapper[4944]: I1124 09:04:37.503302 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz" event={"ID":"299a592e-b90c-436a-9aa4-c2b949cb0e85","Type":"ContainerStarted","Data":"609b818d7be9d935f37746e381aa0211dbd6fafa8cc35ec1afd51e54122a1e13"} Nov 24 09:04:37 crc kubenswrapper[4944]: I1124 09:04:37.506267 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-fdbf99894-8q8pr" event={"ID":"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7","Type":"ContainerStarted","Data":"a5126936fb0e5cb640d06c91f5f506597a25d583bf5a4aa81ade4e5ad9e88b03"} Nov 24 09:04:37 crc kubenswrapper[4944]: I1124 09:04:37.506290 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-fdbf99894-8q8pr" event={"ID":"e6b4d1cd-da30-49de-a548-5e57f4aaa7c7","Type":"ContainerStarted","Data":"89e551b2a8f12d84a3d32d697dea9bc31d82022e564f10937eb5f4958bb297ed"} Nov 24 09:04:37 crc kubenswrapper[4944]: I1124 09:04:37.522841 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-fdbf99894-8q8pr" podStartSLOduration=1.522824231 podStartE2EDuration="1.522824231s" podCreationTimestamp="2025-11-24 09:04:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:04:37.521710287 +0000 UTC m=+738.056150749" watchObservedRunningTime="2025-11-24 09:04:37.522824231 +0000 UTC m=+738.057264693" Nov 24 09:04:38 crc kubenswrapper[4944]: E1124 09:04:38.012147 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice/crio-6b2a0fdd20b2cdb67c1d1c91d42e55e5ce28688508b593997dd8566e4a138d36\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice\": RecentStats: unable to find data in memory cache]" Nov 24 09:04:39 crc kubenswrapper[4944]: I1124 09:04:39.523304 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz" event={"ID":"299a592e-b90c-436a-9aa4-c2b949cb0e85","Type":"ContainerStarted","Data":"f3a0a739b43d1e8f37791bf86777910acd8eb6f76924e14350134e32822b1299"} Nov 24 09:04:39 crc kubenswrapper[4944]: I1124 09:04:39.524389 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vjtqb" event={"ID":"6a5e82c5-f9a6-45ea-b831-36c6573b974e","Type":"ContainerStarted","Data":"0727fe236cf61c87a7a2666c6e1700a6aa1540f2cf5f39649781766cad574ea1"} Nov 24 09:04:39 crc kubenswrapper[4944]: I1124 09:04:39.524499 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:39 crc kubenswrapper[4944]: I1124 09:04:39.526772 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" event={"ID":"3f916409-7217-43dd-a0fe-859fce69591e","Type":"ContainerStarted","Data":"9bdfb28cf9b39ce1fd55885fdf224ba97b325eb9624d395b109ed499a2550c0b"} Nov 24 09:04:39 crc kubenswrapper[4944]: I1124 09:04:39.526926 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" Nov 24 09:04:39 crc kubenswrapper[4944]: I1124 09:04:39.552663 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-vjtqb" podStartSLOduration=1.089223599 podStartE2EDuration="3.552604398s" podCreationTimestamp="2025-11-24 09:04:36 +0000 UTC" firstStartedPulling="2025-11-24 09:04:36.413194499 +0000 UTC m=+736.947634961" lastFinishedPulling="2025-11-24 09:04:38.876575288 +0000 UTC m=+739.411015760" observedRunningTime="2025-11-24 09:04:39.543630963 +0000 UTC m=+740.078071425" watchObservedRunningTime="2025-11-24 09:04:39.552604398 +0000 UTC m=+740.087044890" Nov 24 09:04:39 crc kubenswrapper[4944]: I1124 09:04:39.560033 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" podStartSLOduration=1.390746972 podStartE2EDuration="3.560014674s" podCreationTimestamp="2025-11-24 09:04:36 +0000 UTC" firstStartedPulling="2025-11-24 09:04:36.612590212 +0000 UTC m=+737.147030674" lastFinishedPulling="2025-11-24 09:04:38.781857914 +0000 UTC m=+739.316298376" observedRunningTime="2025-11-24 09:04:39.558005579 +0000 UTC m=+740.092446041" watchObservedRunningTime="2025-11-24 09:04:39.560014674 +0000 UTC m=+740.094455136" Nov 24 09:04:40 crc kubenswrapper[4944]: I1124 09:04:40.536761 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" event={"ID":"1a7b278b-3ae9-4b09-b586-875bb2cac1d6","Type":"ContainerStarted","Data":"b348378006d81adf37ce9ac2d67df5aee8d53a84af32ecc63e959779e5d0d030"} Nov 24 09:04:40 crc kubenswrapper[4944]: I1124 09:04:40.554665 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-74c8z" podStartSLOduration=2.052167319 podStartE2EDuration="4.554645579s" podCreationTimestamp="2025-11-24 09:04:36 +0000 UTC" firstStartedPulling="2025-11-24 09:04:37.306901243 +0000 UTC m=+737.841341705" lastFinishedPulling="2025-11-24 09:04:39.809379503 +0000 UTC m=+740.343819965" observedRunningTime="2025-11-24 09:04:40.550329933 +0000 UTC m=+741.084770415" watchObservedRunningTime="2025-11-24 09:04:40.554645579 +0000 UTC m=+741.089086041" Nov 24 09:04:41 crc kubenswrapper[4944]: I1124 09:04:41.543080 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz" event={"ID":"299a592e-b90c-436a-9aa4-c2b949cb0e85","Type":"ContainerStarted","Data":"df28f5e0a1f7e796af3dd149e1f6700aea4a34e024bde042a14336dbaee82378"} Nov 24 09:04:41 crc kubenswrapper[4944]: I1124 09:04:41.561401 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-pqwqz" podStartSLOduration=1.145149234 podStartE2EDuration="5.56138001s" podCreationTimestamp="2025-11-24 09:04:36 +0000 UTC" firstStartedPulling="2025-11-24 09:04:36.639913729 +0000 UTC m=+737.174354191" lastFinishedPulling="2025-11-24 09:04:41.056144505 +0000 UTC m=+741.590584967" observedRunningTime="2025-11-24 09:04:41.55792476 +0000 UTC m=+742.092365232" watchObservedRunningTime="2025-11-24 09:04:41.56138001 +0000 UTC m=+742.095820472" Nov 24 09:04:46 crc kubenswrapper[4944]: I1124 09:04:46.400793 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-vjtqb" Nov 24 09:04:46 crc kubenswrapper[4944]: I1124 09:04:46.705210 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:46 crc kubenswrapper[4944]: I1124 09:04:46.705272 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:46 crc kubenswrapper[4944]: I1124 09:04:46.710562 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:47 crc kubenswrapper[4944]: I1124 09:04:47.577560 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-fdbf99894-8q8pr" Nov 24 09:04:47 crc kubenswrapper[4944]: I1124 09:04:47.631004 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-s9r2p"] Nov 24 09:04:48 crc kubenswrapper[4944]: E1124 09:04:48.129775 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice/crio-6b2a0fdd20b2cdb67c1d1c91d42e55e5ce28688508b593997dd8566e4a138d36\": RecentStats: unable to find data in memory cache]" Nov 24 09:04:48 crc kubenswrapper[4944]: I1124 09:04:48.612831 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vsszz"] Nov 24 09:04:48 crc kubenswrapper[4944]: I1124 09:04:48.613424 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" podUID="e33b0db2-705e-4424-985c-61d443a4cded" containerName="controller-manager" containerID="cri-o://de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836" gracePeriod=30 Nov 24 09:04:48 crc kubenswrapper[4944]: I1124 09:04:48.723943 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7"] Nov 24 09:04:48 crc kubenswrapper[4944]: I1124 09:04:48.724316 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" podUID="12954baa-8292-40c1-982a-28e23800078e" containerName="route-controller-manager" containerID="cri-o://6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792" gracePeriod=30 Nov 24 09:04:48 crc kubenswrapper[4944]: I1124 09:04:48.993802 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.078857 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.130619 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-config\") pod \"e33b0db2-705e-4424-985c-61d443a4cded\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.131126 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sgvz\" (UniqueName: \"kubernetes.io/projected/e33b0db2-705e-4424-985c-61d443a4cded-kube-api-access-7sgvz\") pod \"e33b0db2-705e-4424-985c-61d443a4cded\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.131149 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-proxy-ca-bundles\") pod \"e33b0db2-705e-4424-985c-61d443a4cded\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.131233 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-client-ca\") pod \"e33b0db2-705e-4424-985c-61d443a4cded\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.131259 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e33b0db2-705e-4424-985c-61d443a4cded-serving-cert\") pod \"e33b0db2-705e-4424-985c-61d443a4cded\" (UID: \"e33b0db2-705e-4424-985c-61d443a4cded\") " Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.131780 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-config" (OuterVolumeSpecName: "config") pod "e33b0db2-705e-4424-985c-61d443a4cded" (UID: "e33b0db2-705e-4424-985c-61d443a4cded"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.131797 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-client-ca" (OuterVolumeSpecName: "client-ca") pod "e33b0db2-705e-4424-985c-61d443a4cded" (UID: "e33b0db2-705e-4424-985c-61d443a4cded"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.131804 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e33b0db2-705e-4424-985c-61d443a4cded" (UID: "e33b0db2-705e-4424-985c-61d443a4cded"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.137104 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e33b0db2-705e-4424-985c-61d443a4cded-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e33b0db2-705e-4424-985c-61d443a4cded" (UID: "e33b0db2-705e-4424-985c-61d443a4cded"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.137175 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e33b0db2-705e-4424-985c-61d443a4cded-kube-api-access-7sgvz" (OuterVolumeSpecName: "kube-api-access-7sgvz") pod "e33b0db2-705e-4424-985c-61d443a4cded" (UID: "e33b0db2-705e-4424-985c-61d443a4cded"). InnerVolumeSpecName "kube-api-access-7sgvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.232121 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12954baa-8292-40c1-982a-28e23800078e-serving-cert\") pod \"12954baa-8292-40c1-982a-28e23800078e\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.232194 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms79n\" (UniqueName: \"kubernetes.io/projected/12954baa-8292-40c1-982a-28e23800078e-kube-api-access-ms79n\") pod \"12954baa-8292-40c1-982a-28e23800078e\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.232227 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-config\") pod \"12954baa-8292-40c1-982a-28e23800078e\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.232253 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-client-ca\") pod \"12954baa-8292-40c1-982a-28e23800078e\" (UID: \"12954baa-8292-40c1-982a-28e23800078e\") " Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.232464 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e33b0db2-705e-4424-985c-61d443a4cded-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.232480 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.232489 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sgvz\" (UniqueName: \"kubernetes.io/projected/e33b0db2-705e-4424-985c-61d443a4cded-kube-api-access-7sgvz\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.232499 4944 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.232507 4944 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e33b0db2-705e-4424-985c-61d443a4cded-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.233358 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-client-ca" (OuterVolumeSpecName: "client-ca") pod "12954baa-8292-40c1-982a-28e23800078e" (UID: "12954baa-8292-40c1-982a-28e23800078e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.233512 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-config" (OuterVolumeSpecName: "config") pod "12954baa-8292-40c1-982a-28e23800078e" (UID: "12954baa-8292-40c1-982a-28e23800078e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.236106 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12954baa-8292-40c1-982a-28e23800078e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "12954baa-8292-40c1-982a-28e23800078e" (UID: "12954baa-8292-40c1-982a-28e23800078e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.237527 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12954baa-8292-40c1-982a-28e23800078e-kube-api-access-ms79n" (OuterVolumeSpecName: "kube-api-access-ms79n") pod "12954baa-8292-40c1-982a-28e23800078e" (UID: "12954baa-8292-40c1-982a-28e23800078e"). InnerVolumeSpecName "kube-api-access-ms79n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.333493 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms79n\" (UniqueName: \"kubernetes.io/projected/12954baa-8292-40c1-982a-28e23800078e-kube-api-access-ms79n\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.333532 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.333540 4944 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12954baa-8292-40c1-982a-28e23800078e-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.333548 4944 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12954baa-8292-40c1-982a-28e23800078e-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.585686 4944 generic.go:334] "Generic (PLEG): container finished" podID="e33b0db2-705e-4424-985c-61d443a4cded" containerID="de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836" exitCode=0 Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.585743 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.585767 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" event={"ID":"e33b0db2-705e-4424-985c-61d443a4cded","Type":"ContainerDied","Data":"de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836"} Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.585796 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vsszz" event={"ID":"e33b0db2-705e-4424-985c-61d443a4cded","Type":"ContainerDied","Data":"cc92af44ae7808071c687f1aebb8b4d0ec50a20afa833f7b18f9044b36139745"} Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.585812 4944 scope.go:117] "RemoveContainer" containerID="de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.587251 4944 generic.go:334] "Generic (PLEG): container finished" podID="12954baa-8292-40c1-982a-28e23800078e" containerID="6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792" exitCode=0 Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.587273 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" event={"ID":"12954baa-8292-40c1-982a-28e23800078e","Type":"ContainerDied","Data":"6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792"} Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.587286 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" event={"ID":"12954baa-8292-40c1-982a-28e23800078e","Type":"ContainerDied","Data":"3007e310db599d5ce9c596a3060d492ffaf7ca900b29c9727de39d4d79f52453"} Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.587329 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.600496 4944 scope.go:117] "RemoveContainer" containerID="de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836" Nov 24 09:04:49 crc kubenswrapper[4944]: E1124 09:04:49.601009 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836\": container with ID starting with de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836 not found: ID does not exist" containerID="de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.601079 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836"} err="failed to get container status \"de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836\": rpc error: code = NotFound desc = could not find container \"de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836\": container with ID starting with de6e0da6399babe7f2979b6d9e4bb2a931f6aaa1593c2e5c444c30737419b836 not found: ID does not exist" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.601116 4944 scope.go:117] "RemoveContainer" containerID="6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.615662 4944 scope.go:117] "RemoveContainer" containerID="6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792" Nov 24 09:04:49 crc kubenswrapper[4944]: E1124 09:04:49.616023 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792\": container with ID starting with 6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792 not found: ID does not exist" containerID="6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.616068 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792"} err="failed to get container status \"6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792\": rpc error: code = NotFound desc = could not find container \"6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792\": container with ID starting with 6152dd85fb05ff015dfde90098c3399d56214a33d9a95acd4a4ecdc2429e8792 not found: ID does not exist" Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.616687 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7"] Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.620685 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v7pc7"] Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.633260 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vsszz"] Nov 24 09:04:49 crc kubenswrapper[4944]: I1124 09:04:49.637007 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vsszz"] Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.286882 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12954baa-8292-40c1-982a-28e23800078e" path="/var/lib/kubelet/pods/12954baa-8292-40c1-982a-28e23800078e/volumes" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.287438 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e33b0db2-705e-4424-985c-61d443a4cded" path="/var/lib/kubelet/pods/e33b0db2-705e-4424-985c-61d443a4cded/volumes" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.637625 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6bb4dff678-7wgpz"] Nov 24 09:04:50 crc kubenswrapper[4944]: E1124 09:04:50.637890 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12954baa-8292-40c1-982a-28e23800078e" containerName="route-controller-manager" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.637906 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="12954baa-8292-40c1-982a-28e23800078e" containerName="route-controller-manager" Nov 24 09:04:50 crc kubenswrapper[4944]: E1124 09:04:50.637919 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e33b0db2-705e-4424-985c-61d443a4cded" containerName="controller-manager" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.637927 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e33b0db2-705e-4424-985c-61d443a4cded" containerName="controller-manager" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.638104 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e33b0db2-705e-4424-985c-61d443a4cded" containerName="controller-manager" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.638121 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="12954baa-8292-40c1-982a-28e23800078e" containerName="route-controller-manager" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.638534 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.642281 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.644214 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.644449 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.644677 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.644809 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.644917 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.653442 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.660182 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh"] Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.661522 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.667709 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bb4dff678-7wgpz"] Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.673690 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.673922 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.674087 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.674209 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.674322 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.674436 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.698117 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh"] Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.757345 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-serving-cert\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.757418 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b16d1abe-da6a-4f9c-a25b-162754553fd2-serving-cert\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.757449 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b16d1abe-da6a-4f9c-a25b-162754553fd2-config\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.757475 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b16d1abe-da6a-4f9c-a25b-162754553fd2-client-ca\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.757541 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx55j\" (UniqueName: \"kubernetes.io/projected/b16d1abe-da6a-4f9c-a25b-162754553fd2-kube-api-access-vx55j\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.757578 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8ghv\" (UniqueName: \"kubernetes.io/projected/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-kube-api-access-f8ghv\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.757602 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-config\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.757647 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b16d1abe-da6a-4f9c-a25b-162754553fd2-proxy-ca-bundles\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.757675 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-client-ca\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.858489 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b16d1abe-da6a-4f9c-a25b-162754553fd2-serving-cert\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.858552 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b16d1abe-da6a-4f9c-a25b-162754553fd2-config\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.858585 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b16d1abe-da6a-4f9c-a25b-162754553fd2-client-ca\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.858626 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx55j\" (UniqueName: \"kubernetes.io/projected/b16d1abe-da6a-4f9c-a25b-162754553fd2-kube-api-access-vx55j\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.858671 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8ghv\" (UniqueName: \"kubernetes.io/projected/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-kube-api-access-f8ghv\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.858697 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-config\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.858731 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b16d1abe-da6a-4f9c-a25b-162754553fd2-proxy-ca-bundles\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.858763 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-client-ca\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.858800 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-serving-cert\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.859856 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b16d1abe-da6a-4f9c-a25b-162754553fd2-client-ca\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.860015 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-client-ca\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.860256 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-config\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.860320 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b16d1abe-da6a-4f9c-a25b-162754553fd2-config\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.860736 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b16d1abe-da6a-4f9c-a25b-162754553fd2-proxy-ca-bundles\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.862415 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-serving-cert\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.862588 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b16d1abe-da6a-4f9c-a25b-162754553fd2-serving-cert\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.874881 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8ghv\" (UniqueName: \"kubernetes.io/projected/d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204-kube-api-access-f8ghv\") pod \"route-controller-manager-7967885c47-j5zzh\" (UID: \"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204\") " pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.875842 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx55j\" (UniqueName: \"kubernetes.io/projected/b16d1abe-da6a-4f9c-a25b-162754553fd2-kube-api-access-vx55j\") pod \"controller-manager-6bb4dff678-7wgpz\" (UID: \"b16d1abe-da6a-4f9c-a25b-162754553fd2\") " pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:50 crc kubenswrapper[4944]: I1124 09:04:50.993118 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.007358 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.212389 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bb4dff678-7wgpz"] Nov 24 09:04:51 crc kubenswrapper[4944]: W1124 09:04:51.217699 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb16d1abe_da6a_4f9c_a25b_162754553fd2.slice/crio-5c2d68514963e18983b65c96d460111a9cd1321da51c2926cc863bf625a27f29 WatchSource:0}: Error finding container 5c2d68514963e18983b65c96d460111a9cd1321da51c2926cc863bf625a27f29: Status 404 returned error can't find the container with id 5c2d68514963e18983b65c96d460111a9cd1321da51c2926cc863bf625a27f29 Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.276626 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh"] Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.600515 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" event={"ID":"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204","Type":"ContainerStarted","Data":"9245191072dfd0fb45149fa2af0abb57b1188dcaeb0924f05ed3433def9c8709"} Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.600843 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" event={"ID":"d0a7fd51-1d08-4c5b-94e0-f4d6bfba1204","Type":"ContainerStarted","Data":"79d06d267e29d8835f2b48e53e222ff15b73e0dec5814bfe7789811c23a09e56"} Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.601922 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.603741 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" event={"ID":"b16d1abe-da6a-4f9c-a25b-162754553fd2","Type":"ContainerStarted","Data":"4b4f21440ddbb470efdccecac669231fd36c0c1cdd8d87c10b48a7dd631ebd03"} Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.603770 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" event={"ID":"b16d1abe-da6a-4f9c-a25b-162754553fd2","Type":"ContainerStarted","Data":"5c2d68514963e18983b65c96d460111a9cd1321da51c2926cc863bf625a27f29"} Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.604329 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.620891 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.628964 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" podStartSLOduration=3.628946365 podStartE2EDuration="3.628946365s" podCreationTimestamp="2025-11-24 09:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:04:51.627869471 +0000 UTC m=+752.162309943" watchObservedRunningTime="2025-11-24 09:04:51.628946365 +0000 UTC m=+752.163386827" Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.656225 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6bb4dff678-7wgpz" podStartSLOduration=3.656204179 podStartE2EDuration="3.656204179s" podCreationTimestamp="2025-11-24 09:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:04:51.652887694 +0000 UTC m=+752.187328186" watchObservedRunningTime="2025-11-24 09:04:51.656204179 +0000 UTC m=+752.190644641" Nov 24 09:04:51 crc kubenswrapper[4944]: I1124 09:04:51.918339 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7967885c47-j5zzh" Nov 24 09:04:53 crc kubenswrapper[4944]: I1124 09:04:53.548339 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:04:53 crc kubenswrapper[4944]: I1124 09:04:53.548972 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:04:53 crc kubenswrapper[4944]: I1124 09:04:53.549078 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:04:53 crc kubenswrapper[4944]: I1124 09:04:53.550133 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f61d46fb34a7fb69e551ab18c2960366dce6b114b94216eab6db96b03be79c6f"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:04:53 crc kubenswrapper[4944]: I1124 09:04:53.550222 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://f61d46fb34a7fb69e551ab18c2960366dce6b114b94216eab6db96b03be79c6f" gracePeriod=600 Nov 24 09:04:54 crc kubenswrapper[4944]: I1124 09:04:54.624400 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="f61d46fb34a7fb69e551ab18c2960366dce6b114b94216eab6db96b03be79c6f" exitCode=0 Nov 24 09:04:54 crc kubenswrapper[4944]: I1124 09:04:54.624483 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"f61d46fb34a7fb69e551ab18c2960366dce6b114b94216eab6db96b03be79c6f"} Nov 24 09:04:54 crc kubenswrapper[4944]: I1124 09:04:54.625956 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"b51cb7736616a4e1de4d8a3f423bb2e041c1d8aa22c653e62b3b1a70f4ce6afd"} Nov 24 09:04:54 crc kubenswrapper[4944]: I1124 09:04:54.626006 4944 scope.go:117] "RemoveContainer" containerID="f05b7d2e929ec321cc8c8fd2068636a82f71fce547e3746a863561b7880eca47" Nov 24 09:04:56 crc kubenswrapper[4944]: I1124 09:04:56.155348 4944 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 09:04:56 crc kubenswrapper[4944]: I1124 09:04:56.362265 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-sxnnb" Nov 24 09:04:58 crc kubenswrapper[4944]: E1124 09:04:58.255089 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice/crio-6b2a0fdd20b2cdb67c1d1c91d42e55e5ce28688508b593997dd8566e4a138d36\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice\": RecentStats: unable to find data in memory cache]" Nov 24 09:05:08 crc kubenswrapper[4944]: E1124 09:05:08.371511 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice/crio-6b2a0fdd20b2cdb67c1d1c91d42e55e5ce28688508b593997dd8566e4a138d36\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice\": RecentStats: unable to find data in memory cache]" Nov 24 09:05:10 crc kubenswrapper[4944]: I1124 09:05:10.828677 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7"] Nov 24 09:05:10 crc kubenswrapper[4944]: I1124 09:05:10.830317 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:10 crc kubenswrapper[4944]: I1124 09:05:10.832632 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 09:05:10 crc kubenswrapper[4944]: I1124 09:05:10.843727 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7"] Nov 24 09:05:10 crc kubenswrapper[4944]: I1124 09:05:10.957042 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mk27\" (UniqueName: \"kubernetes.io/projected/3fbed182-a3a3-46f3-b6af-6383550b36e1-kube-api-access-4mk27\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:10 crc kubenswrapper[4944]: I1124 09:05:10.957200 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:10 crc kubenswrapper[4944]: I1124 09:05:10.957431 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:11 crc kubenswrapper[4944]: I1124 09:05:11.058459 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:11 crc kubenswrapper[4944]: I1124 09:05:11.058529 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mk27\" (UniqueName: \"kubernetes.io/projected/3fbed182-a3a3-46f3-b6af-6383550b36e1-kube-api-access-4mk27\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:11 crc kubenswrapper[4944]: I1124 09:05:11.058566 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:11 crc kubenswrapper[4944]: I1124 09:05:11.059106 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:11 crc kubenswrapper[4944]: I1124 09:05:11.059133 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:11 crc kubenswrapper[4944]: I1124 09:05:11.076198 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mk27\" (UniqueName: \"kubernetes.io/projected/3fbed182-a3a3-46f3-b6af-6383550b36e1-kube-api-access-4mk27\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:11 crc kubenswrapper[4944]: I1124 09:05:11.151118 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:11 crc kubenswrapper[4944]: I1124 09:05:11.534080 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7"] Nov 24 09:05:11 crc kubenswrapper[4944]: I1124 09:05:11.726487 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" event={"ID":"3fbed182-a3a3-46f3-b6af-6383550b36e1","Type":"ContainerStarted","Data":"8b1d21db1c92f193c4431f5b94eb3221729a1ec81583733bfccf0934237f9376"} Nov 24 09:05:11 crc kubenswrapper[4944]: I1124 09:05:11.726800 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" event={"ID":"3fbed182-a3a3-46f3-b6af-6383550b36e1","Type":"ContainerStarted","Data":"52508c8bf81ba66eae3c1f374481c247493ca915922e5fbe5146e2bc19c53412"} Nov 24 09:05:12 crc kubenswrapper[4944]: I1124 09:05:12.677918 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-s9r2p" podUID="b355df00-f706-42c5-b16f-8c237a5048d8" containerName="console" containerID="cri-o://44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2" gracePeriod=15 Nov 24 09:05:12 crc kubenswrapper[4944]: I1124 09:05:12.732829 4944 generic.go:334] "Generic (PLEG): container finished" podID="3fbed182-a3a3-46f3-b6af-6383550b36e1" containerID="8b1d21db1c92f193c4431f5b94eb3221729a1ec81583733bfccf0934237f9376" exitCode=0 Nov 24 09:05:12 crc kubenswrapper[4944]: I1124 09:05:12.732871 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" event={"ID":"3fbed182-a3a3-46f3-b6af-6383550b36e1","Type":"ContainerDied","Data":"8b1d21db1c92f193c4431f5b94eb3221729a1ec81583733bfccf0934237f9376"} Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.126902 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-s9r2p_b355df00-f706-42c5-b16f-8c237a5048d8/console/0.log" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.127259 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.185302 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-serving-cert\") pod \"b355df00-f706-42c5-b16f-8c237a5048d8\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.185364 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-trusted-ca-bundle\") pod \"b355df00-f706-42c5-b16f-8c237a5048d8\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.185397 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-oauth-serving-cert\") pod \"b355df00-f706-42c5-b16f-8c237a5048d8\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.185472 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-console-config\") pod \"b355df00-f706-42c5-b16f-8c237a5048d8\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.185530 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-service-ca\") pod \"b355df00-f706-42c5-b16f-8c237a5048d8\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.185566 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp7xs\" (UniqueName: \"kubernetes.io/projected/b355df00-f706-42c5-b16f-8c237a5048d8-kube-api-access-kp7xs\") pod \"b355df00-f706-42c5-b16f-8c237a5048d8\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.185596 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-oauth-config\") pod \"b355df00-f706-42c5-b16f-8c237a5048d8\" (UID: \"b355df00-f706-42c5-b16f-8c237a5048d8\") " Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.186893 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b355df00-f706-42c5-b16f-8c237a5048d8" (UID: "b355df00-f706-42c5-b16f-8c237a5048d8"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.186916 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-service-ca" (OuterVolumeSpecName: "service-ca") pod "b355df00-f706-42c5-b16f-8c237a5048d8" (UID: "b355df00-f706-42c5-b16f-8c237a5048d8"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.186923 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b355df00-f706-42c5-b16f-8c237a5048d8" (UID: "b355df00-f706-42c5-b16f-8c237a5048d8"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.187172 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-console-config" (OuterVolumeSpecName: "console-config") pod "b355df00-f706-42c5-b16f-8c237a5048d8" (UID: "b355df00-f706-42c5-b16f-8c237a5048d8"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.191960 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b355df00-f706-42c5-b16f-8c237a5048d8" (UID: "b355df00-f706-42c5-b16f-8c237a5048d8"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.192400 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b355df00-f706-42c5-b16f-8c237a5048d8-kube-api-access-kp7xs" (OuterVolumeSpecName: "kube-api-access-kp7xs") pod "b355df00-f706-42c5-b16f-8c237a5048d8" (UID: "b355df00-f706-42c5-b16f-8c237a5048d8"). InnerVolumeSpecName "kube-api-access-kp7xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.192545 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b355df00-f706-42c5-b16f-8c237a5048d8" (UID: "b355df00-f706-42c5-b16f-8c237a5048d8"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.286567 4944 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.286630 4944 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.286641 4944 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.286651 4944 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.286659 4944 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b355df00-f706-42c5-b16f-8c237a5048d8-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.286667 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp7xs\" (UniqueName: \"kubernetes.io/projected/b355df00-f706-42c5-b16f-8c237a5048d8-kube-api-access-kp7xs\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.286675 4944 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b355df00-f706-42c5-b16f-8c237a5048d8-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.743182 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-s9r2p_b355df00-f706-42c5-b16f-8c237a5048d8/console/0.log" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.743231 4944 generic.go:334] "Generic (PLEG): container finished" podID="b355df00-f706-42c5-b16f-8c237a5048d8" containerID="44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2" exitCode=2 Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.743307 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s9r2p" event={"ID":"b355df00-f706-42c5-b16f-8c237a5048d8","Type":"ContainerDied","Data":"44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2"} Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.743332 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-s9r2p" event={"ID":"b355df00-f706-42c5-b16f-8c237a5048d8","Type":"ContainerDied","Data":"4782a3a25e0ad3ac5237b4047d6456430bc1b17cfee15ca912c0935dfd784881"} Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.743348 4944 scope.go:117] "RemoveContainer" containerID="44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.743446 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-s9r2p" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.747177 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" event={"ID":"3fbed182-a3a3-46f3-b6af-6383550b36e1","Type":"ContainerStarted","Data":"55966f3f3de1e2db211434104bb7a5f80d34e761323a5f8d3bab15f336582d02"} Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.770341 4944 scope.go:117] "RemoveContainer" containerID="44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2" Nov 24 09:05:13 crc kubenswrapper[4944]: E1124 09:05:13.770774 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2\": container with ID starting with 44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2 not found: ID does not exist" containerID="44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.770815 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2"} err="failed to get container status \"44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2\": rpc error: code = NotFound desc = could not find container \"44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2\": container with ID starting with 44f3c68dbd5f102137fc633f0c609da3ec73d73ee61859b49d4e7f41acb1b8e2 not found: ID does not exist" Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.782143 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-s9r2p"] Nov 24 09:05:13 crc kubenswrapper[4944]: I1124 09:05:13.785821 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-s9r2p"] Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.195860 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z6pz7"] Nov 24 09:05:14 crc kubenswrapper[4944]: E1124 09:05:14.196262 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b355df00-f706-42c5-b16f-8c237a5048d8" containerName="console" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.196284 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b355df00-f706-42c5-b16f-8c237a5048d8" containerName="console" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.196411 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b355df00-f706-42c5-b16f-8c237a5048d8" containerName="console" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.203700 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6pz7"] Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.203885 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.284073 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b355df00-f706-42c5-b16f-8c237a5048d8" path="/var/lib/kubelet/pods/b355df00-f706-42c5-b16f-8c237a5048d8/volumes" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.300665 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-utilities\") pod \"redhat-operators-z6pz7\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.300761 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-catalog-content\") pod \"redhat-operators-z6pz7\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.300791 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnl56\" (UniqueName: \"kubernetes.io/projected/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-kube-api-access-qnl56\") pod \"redhat-operators-z6pz7\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.402374 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-utilities\") pod \"redhat-operators-z6pz7\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.402554 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-catalog-content\") pod \"redhat-operators-z6pz7\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.402598 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnl56\" (UniqueName: \"kubernetes.io/projected/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-kube-api-access-qnl56\") pod \"redhat-operators-z6pz7\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.403241 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-utilities\") pod \"redhat-operators-z6pz7\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.403506 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-catalog-content\") pod \"redhat-operators-z6pz7\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.422659 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnl56\" (UniqueName: \"kubernetes.io/projected/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-kube-api-access-qnl56\") pod \"redhat-operators-z6pz7\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.533175 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.760502 4944 generic.go:334] "Generic (PLEG): container finished" podID="3fbed182-a3a3-46f3-b6af-6383550b36e1" containerID="55966f3f3de1e2db211434104bb7a5f80d34e761323a5f8d3bab15f336582d02" exitCode=0 Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.760580 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" event={"ID":"3fbed182-a3a3-46f3-b6af-6383550b36e1","Type":"ContainerDied","Data":"55966f3f3de1e2db211434104bb7a5f80d34e761323a5f8d3bab15f336582d02"} Nov 24 09:05:14 crc kubenswrapper[4944]: I1124 09:05:14.956797 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z6pz7"] Nov 24 09:05:14 crc kubenswrapper[4944]: W1124 09:05:14.970249 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a4aef39_55c4_46d1_9c3f_0f84c4a25fcd.slice/crio-8b20f88c54faa1057841dad22eccc30b396151d2d0965b40adc11d733b6ae10d WatchSource:0}: Error finding container 8b20f88c54faa1057841dad22eccc30b396151d2d0965b40adc11d733b6ae10d: Status 404 returned error can't find the container with id 8b20f88c54faa1057841dad22eccc30b396151d2d0965b40adc11d733b6ae10d Nov 24 09:05:15 crc kubenswrapper[4944]: I1124 09:05:15.784000 4944 generic.go:334] "Generic (PLEG): container finished" podID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" containerID="7cac9fc602db7e81d1e4a3780468fee39d1d32a49e148d2f5c40ef0d558518d5" exitCode=0 Nov 24 09:05:15 crc kubenswrapper[4944]: I1124 09:05:15.784113 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6pz7" event={"ID":"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd","Type":"ContainerDied","Data":"7cac9fc602db7e81d1e4a3780468fee39d1d32a49e148d2f5c40ef0d558518d5"} Nov 24 09:05:15 crc kubenswrapper[4944]: I1124 09:05:15.784142 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6pz7" event={"ID":"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd","Type":"ContainerStarted","Data":"8b20f88c54faa1057841dad22eccc30b396151d2d0965b40adc11d733b6ae10d"} Nov 24 09:05:15 crc kubenswrapper[4944]: I1124 09:05:15.786026 4944 generic.go:334] "Generic (PLEG): container finished" podID="3fbed182-a3a3-46f3-b6af-6383550b36e1" containerID="26cbaca4ca7a8266ebe659d06bfd698feff43cf5c6f65f99ca2e209eaa595499" exitCode=0 Nov 24 09:05:15 crc kubenswrapper[4944]: I1124 09:05:15.786062 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" event={"ID":"3fbed182-a3a3-46f3-b6af-6383550b36e1","Type":"ContainerDied","Data":"26cbaca4ca7a8266ebe659d06bfd698feff43cf5c6f65f99ca2e209eaa595499"} Nov 24 09:05:16 crc kubenswrapper[4944]: I1124 09:05:16.793878 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6pz7" event={"ID":"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd","Type":"ContainerStarted","Data":"266d6af880b384a8c421cb7e3d054d4eb33f887317423c4452f46e93d3dcfa81"} Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.127869 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.241244 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-bundle\") pod \"3fbed182-a3a3-46f3-b6af-6383550b36e1\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.241435 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mk27\" (UniqueName: \"kubernetes.io/projected/3fbed182-a3a3-46f3-b6af-6383550b36e1-kube-api-access-4mk27\") pod \"3fbed182-a3a3-46f3-b6af-6383550b36e1\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.241458 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-util\") pod \"3fbed182-a3a3-46f3-b6af-6383550b36e1\" (UID: \"3fbed182-a3a3-46f3-b6af-6383550b36e1\") " Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.242422 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-bundle" (OuterVolumeSpecName: "bundle") pod "3fbed182-a3a3-46f3-b6af-6383550b36e1" (UID: "3fbed182-a3a3-46f3-b6af-6383550b36e1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.248948 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fbed182-a3a3-46f3-b6af-6383550b36e1-kube-api-access-4mk27" (OuterVolumeSpecName: "kube-api-access-4mk27") pod "3fbed182-a3a3-46f3-b6af-6383550b36e1" (UID: "3fbed182-a3a3-46f3-b6af-6383550b36e1"). InnerVolumeSpecName "kube-api-access-4mk27". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.328635 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-util" (OuterVolumeSpecName: "util") pod "3fbed182-a3a3-46f3-b6af-6383550b36e1" (UID: "3fbed182-a3a3-46f3-b6af-6383550b36e1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.344690 4944 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.344806 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mk27\" (UniqueName: \"kubernetes.io/projected/3fbed182-a3a3-46f3-b6af-6383550b36e1-kube-api-access-4mk27\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.344837 4944 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3fbed182-a3a3-46f3-b6af-6383550b36e1-util\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.802360 4944 generic.go:334] "Generic (PLEG): container finished" podID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" containerID="266d6af880b384a8c421cb7e3d054d4eb33f887317423c4452f46e93d3dcfa81" exitCode=0 Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.802464 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6pz7" event={"ID":"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd","Type":"ContainerDied","Data":"266d6af880b384a8c421cb7e3d054d4eb33f887317423c4452f46e93d3dcfa81"} Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.804910 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" event={"ID":"3fbed182-a3a3-46f3-b6af-6383550b36e1","Type":"ContainerDied","Data":"52508c8bf81ba66eae3c1f374481c247493ca915922e5fbe5146e2bc19c53412"} Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.804962 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7" Nov 24 09:05:17 crc kubenswrapper[4944]: I1124 09:05:17.804965 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52508c8bf81ba66eae3c1f374481c247493ca915922e5fbe5146e2bc19c53412" Nov 24 09:05:18 crc kubenswrapper[4944]: E1124 09:05:18.541804 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66b08e16_cf20_4ebb_a6b0_e33bd5947b75.slice/crio-6b2a0fdd20b2cdb67c1d1c91d42e55e5ce28688508b593997dd8566e4a138d36\": RecentStats: unable to find data in memory cache]" Nov 24 09:05:18 crc kubenswrapper[4944]: I1124 09:05:18.813940 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6pz7" event={"ID":"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd","Type":"ContainerStarted","Data":"415c9b04774cc93a7b8ef219d08fb44d201d188ae13ed92501c4a1aec5e38f29"} Nov 24 09:05:18 crc kubenswrapper[4944]: I1124 09:05:18.839623 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z6pz7" podStartSLOduration=2.356251522 podStartE2EDuration="4.839604075s" podCreationTimestamp="2025-11-24 09:05:14 +0000 UTC" firstStartedPulling="2025-11-24 09:05:15.786624196 +0000 UTC m=+776.321064658" lastFinishedPulling="2025-11-24 09:05:18.269976749 +0000 UTC m=+778.804417211" observedRunningTime="2025-11-24 09:05:18.835118083 +0000 UTC m=+779.369558565" watchObservedRunningTime="2025-11-24 09:05:18.839604075 +0000 UTC m=+779.374044537" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.193147 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kx9ww"] Nov 24 09:05:24 crc kubenswrapper[4944]: E1124 09:05:24.193862 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fbed182-a3a3-46f3-b6af-6383550b36e1" containerName="pull" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.193876 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fbed182-a3a3-46f3-b6af-6383550b36e1" containerName="pull" Nov 24 09:05:24 crc kubenswrapper[4944]: E1124 09:05:24.193887 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fbed182-a3a3-46f3-b6af-6383550b36e1" containerName="extract" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.193894 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fbed182-a3a3-46f3-b6af-6383550b36e1" containerName="extract" Nov 24 09:05:24 crc kubenswrapper[4944]: E1124 09:05:24.193905 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fbed182-a3a3-46f3-b6af-6383550b36e1" containerName="util" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.193910 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fbed182-a3a3-46f3-b6af-6383550b36e1" containerName="util" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.194021 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fbed182-a3a3-46f3-b6af-6383550b36e1" containerName="extract" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.194773 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.207769 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx9ww"] Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.311319 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj"] Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.312084 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.314795 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.315123 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.315256 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-w55ts" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.315388 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.315530 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.327380 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj"] Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.334974 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-catalog-content\") pod \"redhat-marketplace-kx9ww\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.335020 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqhvv\" (UniqueName: \"kubernetes.io/projected/62372b3b-e2f5-40e6-835c-eb3405c22cfb-kube-api-access-jqhvv\") pod \"redhat-marketplace-kx9ww\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.335098 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-utilities\") pod \"redhat-marketplace-kx9ww\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.436886 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-catalog-content\") pod \"redhat-marketplace-kx9ww\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.436941 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqhvv\" (UniqueName: \"kubernetes.io/projected/62372b3b-e2f5-40e6-835c-eb3405c22cfb-kube-api-access-jqhvv\") pod \"redhat-marketplace-kx9ww\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.437011 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-utilities\") pod \"redhat-marketplace-kx9ww\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.437417 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-catalog-content\") pod \"redhat-marketplace-kx9ww\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.437421 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/27694069-e435-46a6-9fcc-b848a4a56fde-webhook-cert\") pod \"metallb-operator-controller-manager-6c8c4f86d5-h8tbj\" (UID: \"27694069-e435-46a6-9fcc-b848a4a56fde\") " pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.437500 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-utilities\") pod \"redhat-marketplace-kx9ww\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.437585 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68szk\" (UniqueName: \"kubernetes.io/projected/27694069-e435-46a6-9fcc-b848a4a56fde-kube-api-access-68szk\") pod \"metallb-operator-controller-manager-6c8c4f86d5-h8tbj\" (UID: \"27694069-e435-46a6-9fcc-b848a4a56fde\") " pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.437626 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/27694069-e435-46a6-9fcc-b848a4a56fde-apiservice-cert\") pod \"metallb-operator-controller-manager-6c8c4f86d5-h8tbj\" (UID: \"27694069-e435-46a6-9fcc-b848a4a56fde\") " pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.462964 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqhvv\" (UniqueName: \"kubernetes.io/projected/62372b3b-e2f5-40e6-835c-eb3405c22cfb-kube-api-access-jqhvv\") pod \"redhat-marketplace-kx9ww\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.512134 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.533719 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.534451 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.538567 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68szk\" (UniqueName: \"kubernetes.io/projected/27694069-e435-46a6-9fcc-b848a4a56fde-kube-api-access-68szk\") pod \"metallb-operator-controller-manager-6c8c4f86d5-h8tbj\" (UID: \"27694069-e435-46a6-9fcc-b848a4a56fde\") " pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.538626 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/27694069-e435-46a6-9fcc-b848a4a56fde-apiservice-cert\") pod \"metallb-operator-controller-manager-6c8c4f86d5-h8tbj\" (UID: \"27694069-e435-46a6-9fcc-b848a4a56fde\") " pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.538712 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/27694069-e435-46a6-9fcc-b848a4a56fde-webhook-cert\") pod \"metallb-operator-controller-manager-6c8c4f86d5-h8tbj\" (UID: \"27694069-e435-46a6-9fcc-b848a4a56fde\") " pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.546080 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/27694069-e435-46a6-9fcc-b848a4a56fde-webhook-cert\") pod \"metallb-operator-controller-manager-6c8c4f86d5-h8tbj\" (UID: \"27694069-e435-46a6-9fcc-b848a4a56fde\") " pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.547433 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/27694069-e435-46a6-9fcc-b848a4a56fde-apiservice-cert\") pod \"metallb-operator-controller-manager-6c8c4f86d5-h8tbj\" (UID: \"27694069-e435-46a6-9fcc-b848a4a56fde\") " pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.560676 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68szk\" (UniqueName: \"kubernetes.io/projected/27694069-e435-46a6-9fcc-b848a4a56fde-kube-api-access-68szk\") pod \"metallb-operator-controller-manager-6c8c4f86d5-h8tbj\" (UID: \"27694069-e435-46a6-9fcc-b848a4a56fde\") " pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.594354 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.628834 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.855518 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx"] Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.858746 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.862393 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.863368 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-z2pgp" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.863612 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.944835 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.946496 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4dc5d368-9a0c-418e-981e-bde67f08e687-webhook-cert\") pod \"metallb-operator-webhook-server-66c67cfc8f-mnxqx\" (UID: \"4dc5d368-9a0c-418e-981e-bde67f08e687\") " pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.946550 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdkbf\" (UniqueName: \"kubernetes.io/projected/4dc5d368-9a0c-418e-981e-bde67f08e687-kube-api-access-jdkbf\") pod \"metallb-operator-webhook-server-66c67cfc8f-mnxqx\" (UID: \"4dc5d368-9a0c-418e-981e-bde67f08e687\") " pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.946587 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4dc5d368-9a0c-418e-981e-bde67f08e687-apiservice-cert\") pod \"metallb-operator-webhook-server-66c67cfc8f-mnxqx\" (UID: \"4dc5d368-9a0c-418e-981e-bde67f08e687\") " pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:24 crc kubenswrapper[4944]: I1124 09:05:24.954239 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx"] Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.047534 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdkbf\" (UniqueName: \"kubernetes.io/projected/4dc5d368-9a0c-418e-981e-bde67f08e687-kube-api-access-jdkbf\") pod \"metallb-operator-webhook-server-66c67cfc8f-mnxqx\" (UID: \"4dc5d368-9a0c-418e-981e-bde67f08e687\") " pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.047600 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4dc5d368-9a0c-418e-981e-bde67f08e687-apiservice-cert\") pod \"metallb-operator-webhook-server-66c67cfc8f-mnxqx\" (UID: \"4dc5d368-9a0c-418e-981e-bde67f08e687\") " pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.047692 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4dc5d368-9a0c-418e-981e-bde67f08e687-webhook-cert\") pod \"metallb-operator-webhook-server-66c67cfc8f-mnxqx\" (UID: \"4dc5d368-9a0c-418e-981e-bde67f08e687\") " pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.048176 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx9ww"] Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.062652 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4dc5d368-9a0c-418e-981e-bde67f08e687-apiservice-cert\") pod \"metallb-operator-webhook-server-66c67cfc8f-mnxqx\" (UID: \"4dc5d368-9a0c-418e-981e-bde67f08e687\") " pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.062669 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4dc5d368-9a0c-418e-981e-bde67f08e687-webhook-cert\") pod \"metallb-operator-webhook-server-66c67cfc8f-mnxqx\" (UID: \"4dc5d368-9a0c-418e-981e-bde67f08e687\") " pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.074865 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdkbf\" (UniqueName: \"kubernetes.io/projected/4dc5d368-9a0c-418e-981e-bde67f08e687-kube-api-access-jdkbf\") pod \"metallb-operator-webhook-server-66c67cfc8f-mnxqx\" (UID: \"4dc5d368-9a0c-418e-981e-bde67f08e687\") " pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:25 crc kubenswrapper[4944]: W1124 09:05:25.077336 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62372b3b_e2f5_40e6_835c_eb3405c22cfb.slice/crio-b146d34ad1e6ee8f445c427c8c3443f91b3319473e2fc84667eb07f1f7122b85 WatchSource:0}: Error finding container b146d34ad1e6ee8f445c427c8c3443f91b3319473e2fc84667eb07f1f7122b85: Status 404 returned error can't find the container with id b146d34ad1e6ee8f445c427c8c3443f91b3319473e2fc84667eb07f1f7122b85 Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.195268 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m2hr9"] Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.196738 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.197253 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.213922 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m2hr9"] Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.250452 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-catalog-content\") pod \"certified-operators-m2hr9\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.250722 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68jcx\" (UniqueName: \"kubernetes.io/projected/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-kube-api-access-68jcx\") pod \"certified-operators-m2hr9\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.250828 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-utilities\") pod \"certified-operators-m2hr9\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.352298 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-catalog-content\") pod \"certified-operators-m2hr9\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.352363 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68jcx\" (UniqueName: \"kubernetes.io/projected/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-kube-api-access-68jcx\") pod \"certified-operators-m2hr9\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.352426 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-utilities\") pod \"certified-operators-m2hr9\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.353142 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-catalog-content\") pod \"certified-operators-m2hr9\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.353169 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-utilities\") pod \"certified-operators-m2hr9\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.361915 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj"] Nov 24 09:05:25 crc kubenswrapper[4944]: W1124 09:05:25.371212 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27694069_e435_46a6_9fcc_b848a4a56fde.slice/crio-4ab8dc5dc31d6c6297f695cb3e970524fbee680edcb10d1da1fb5e90fb9a2e50 WatchSource:0}: Error finding container 4ab8dc5dc31d6c6297f695cb3e970524fbee680edcb10d1da1fb5e90fb9a2e50: Status 404 returned error can't find the container with id 4ab8dc5dc31d6c6297f695cb3e970524fbee680edcb10d1da1fb5e90fb9a2e50 Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.377074 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68jcx\" (UniqueName: \"kubernetes.io/projected/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-kube-api-access-68jcx\") pod \"certified-operators-m2hr9\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.518327 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.723151 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx"] Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.865635 4944 generic.go:334] "Generic (PLEG): container finished" podID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" containerID="e3ebcd72da6af2ed9a16a466fba8394ec04c25746bca391860f59a7c0d278551" exitCode=0 Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.865725 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx9ww" event={"ID":"62372b3b-e2f5-40e6-835c-eb3405c22cfb","Type":"ContainerDied","Data":"e3ebcd72da6af2ed9a16a466fba8394ec04c25746bca391860f59a7c0d278551"} Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.865758 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx9ww" event={"ID":"62372b3b-e2f5-40e6-835c-eb3405c22cfb","Type":"ContainerStarted","Data":"b146d34ad1e6ee8f445c427c8c3443f91b3319473e2fc84667eb07f1f7122b85"} Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.867825 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" event={"ID":"4dc5d368-9a0c-418e-981e-bde67f08e687","Type":"ContainerStarted","Data":"947f791e455269e1d49768fe3afc41b90949c7ab0f5a31146e1aaa7efa4a323f"} Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.872168 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" event={"ID":"27694069-e435-46a6-9fcc-b848a4a56fde","Type":"ContainerStarted","Data":"4ab8dc5dc31d6c6297f695cb3e970524fbee680edcb10d1da1fb5e90fb9a2e50"} Nov 24 09:05:25 crc kubenswrapper[4944]: I1124 09:05:25.985969 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m2hr9"] Nov 24 09:05:25 crc kubenswrapper[4944]: W1124 09:05:25.990823 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5ffe08c_6b4b_41b1_8407_dc51da72ccb3.slice/crio-f1e13923cadcdadbbaa8040f4235bc0d599211c81bf6bc7ab32de10d8edb1fb9 WatchSource:0}: Error finding container f1e13923cadcdadbbaa8040f4235bc0d599211c81bf6bc7ab32de10d8edb1fb9: Status 404 returned error can't find the container with id f1e13923cadcdadbbaa8040f4235bc0d599211c81bf6bc7ab32de10d8edb1fb9 Nov 24 09:05:26 crc kubenswrapper[4944]: I1124 09:05:26.889507 4944 generic.go:334] "Generic (PLEG): container finished" podID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" containerID="560c1710dfc67f525235706acaee312ac6365c0bbc72ab09d34378aba4223d16" exitCode=0 Nov 24 09:05:26 crc kubenswrapper[4944]: I1124 09:05:26.889628 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hr9" event={"ID":"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3","Type":"ContainerDied","Data":"560c1710dfc67f525235706acaee312ac6365c0bbc72ab09d34378aba4223d16"} Nov 24 09:05:26 crc kubenswrapper[4944]: I1124 09:05:26.890583 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hr9" event={"ID":"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3","Type":"ContainerStarted","Data":"f1e13923cadcdadbbaa8040f4235bc0d599211c81bf6bc7ab32de10d8edb1fb9"} Nov 24 09:05:27 crc kubenswrapper[4944]: I1124 09:05:27.903689 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hr9" event={"ID":"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3","Type":"ContainerStarted","Data":"bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7"} Nov 24 09:05:27 crc kubenswrapper[4944]: I1124 09:05:27.908336 4944 generic.go:334] "Generic (PLEG): container finished" podID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" containerID="78ee99ae4ae67ceb1bfc3a11b4ae9e490fce7f040ff8d2fd0699413da6b5186e" exitCode=0 Nov 24 09:05:27 crc kubenswrapper[4944]: I1124 09:05:27.908404 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx9ww" event={"ID":"62372b3b-e2f5-40e6-835c-eb3405c22cfb","Type":"ContainerDied","Data":"78ee99ae4ae67ceb1bfc3a11b4ae9e490fce7f040ff8d2fd0699413da6b5186e"} Nov 24 09:05:27 crc kubenswrapper[4944]: I1124 09:05:27.992544 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z6pz7"] Nov 24 09:05:27 crc kubenswrapper[4944]: I1124 09:05:27.992744 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z6pz7" podUID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" containerName="registry-server" containerID="cri-o://415c9b04774cc93a7b8ef219d08fb44d201d188ae13ed92501c4a1aec5e38f29" gracePeriod=2 Nov 24 09:05:29 crc kubenswrapper[4944]: I1124 09:05:29.922097 4944 generic.go:334] "Generic (PLEG): container finished" podID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" containerID="415c9b04774cc93a7b8ef219d08fb44d201d188ae13ed92501c4a1aec5e38f29" exitCode=0 Nov 24 09:05:29 crc kubenswrapper[4944]: I1124 09:05:29.922421 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6pz7" event={"ID":"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd","Type":"ContainerDied","Data":"415c9b04774cc93a7b8ef219d08fb44d201d188ae13ed92501c4a1aec5e38f29"} Nov 24 09:05:29 crc kubenswrapper[4944]: I1124 09:05:29.926229 4944 generic.go:334] "Generic (PLEG): container finished" podID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" containerID="bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7" exitCode=0 Nov 24 09:05:29 crc kubenswrapper[4944]: I1124 09:05:29.926267 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hr9" event={"ID":"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3","Type":"ContainerDied","Data":"bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7"} Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.297776 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.333021 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnl56\" (UniqueName: \"kubernetes.io/projected/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-kube-api-access-qnl56\") pod \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.333108 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-utilities\") pod \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.333218 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-catalog-content\") pod \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\" (UID: \"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd\") " Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.333984 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-utilities" (OuterVolumeSpecName: "utilities") pod "8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" (UID: "8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.340648 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-kube-api-access-qnl56" (OuterVolumeSpecName: "kube-api-access-qnl56") pod "8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" (UID: "8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd"). InnerVolumeSpecName "kube-api-access-qnl56". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.434885 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnl56\" (UniqueName: \"kubernetes.io/projected/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-kube-api-access-qnl56\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.434943 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.451020 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" (UID: "8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.536947 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.946532 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z6pz7" event={"ID":"8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd","Type":"ContainerDied","Data":"8b20f88c54faa1057841dad22eccc30b396151d2d0965b40adc11d733b6ae10d"} Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.946832 4944 scope.go:117] "RemoveContainer" containerID="415c9b04774cc93a7b8ef219d08fb44d201d188ae13ed92501c4a1aec5e38f29" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.946598 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z6pz7" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.948862 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" event={"ID":"4dc5d368-9a0c-418e-981e-bde67f08e687","Type":"ContainerStarted","Data":"b70b449df2151cdff6b65608ae69782301651c548ea96440b9e038ab8574bd6c"} Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.949037 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.950427 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" event={"ID":"27694069-e435-46a6-9fcc-b848a4a56fde","Type":"ContainerStarted","Data":"357f23d05528776121edbc6f29fdb2be3bc77152e98da819b03269df48bf19b0"} Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.950608 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.952915 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hr9" event={"ID":"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3","Type":"ContainerStarted","Data":"5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab"} Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.956631 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx9ww" event={"ID":"62372b3b-e2f5-40e6-835c-eb3405c22cfb","Type":"ContainerStarted","Data":"cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b"} Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.969767 4944 scope.go:117] "RemoveContainer" containerID="266d6af880b384a8c421cb7e3d054d4eb33f887317423c4452f46e93d3dcfa81" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.972202 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" podStartSLOduration=1.9775594760000001 podStartE2EDuration="7.9721871s" podCreationTimestamp="2025-11-24 09:05:24 +0000 UTC" firstStartedPulling="2025-11-24 09:05:25.374521157 +0000 UTC m=+785.908961619" lastFinishedPulling="2025-11-24 09:05:31.369148781 +0000 UTC m=+791.903589243" observedRunningTime="2025-11-24 09:05:31.970833847 +0000 UTC m=+792.505274309" watchObservedRunningTime="2025-11-24 09:05:31.9721871 +0000 UTC m=+792.506627562" Nov 24 09:05:31 crc kubenswrapper[4944]: I1124 09:05:31.992558 4944 scope.go:117] "RemoveContainer" containerID="7cac9fc602db7e81d1e4a3780468fee39d1d32a49e148d2f5c40ef0d558518d5" Nov 24 09:05:32 crc kubenswrapper[4944]: I1124 09:05:32.003920 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" podStartSLOduration=2.373357318 podStartE2EDuration="8.003890789s" podCreationTimestamp="2025-11-24 09:05:24 +0000 UTC" firstStartedPulling="2025-11-24 09:05:25.736397419 +0000 UTC m=+786.270837881" lastFinishedPulling="2025-11-24 09:05:31.36693089 +0000 UTC m=+791.901371352" observedRunningTime="2025-11-24 09:05:31.995892055 +0000 UTC m=+792.530332527" watchObservedRunningTime="2025-11-24 09:05:32.003890789 +0000 UTC m=+792.538331251" Nov 24 09:05:32 crc kubenswrapper[4944]: I1124 09:05:32.030230 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kx9ww" podStartSLOduration=2.529284062 podStartE2EDuration="8.030203496s" podCreationTimestamp="2025-11-24 09:05:24 +0000 UTC" firstStartedPulling="2025-11-24 09:05:25.867308417 +0000 UTC m=+786.401748889" lastFinishedPulling="2025-11-24 09:05:31.368227861 +0000 UTC m=+791.902668323" observedRunningTime="2025-11-24 09:05:32.025868919 +0000 UTC m=+792.560309381" watchObservedRunningTime="2025-11-24 09:05:32.030203496 +0000 UTC m=+792.564643958" Nov 24 09:05:32 crc kubenswrapper[4944]: I1124 09:05:32.062885 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m2hr9" podStartSLOduration=2.315145262 podStartE2EDuration="7.062861487s" podCreationTimestamp="2025-11-24 09:05:25 +0000 UTC" firstStartedPulling="2025-11-24 09:05:26.899523979 +0000 UTC m=+787.433964441" lastFinishedPulling="2025-11-24 09:05:31.647240204 +0000 UTC m=+792.181680666" observedRunningTime="2025-11-24 09:05:32.044377698 +0000 UTC m=+792.578818160" watchObservedRunningTime="2025-11-24 09:05:32.062861487 +0000 UTC m=+792.597301949" Nov 24 09:05:32 crc kubenswrapper[4944]: I1124 09:05:32.063971 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z6pz7"] Nov 24 09:05:32 crc kubenswrapper[4944]: I1124 09:05:32.072642 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z6pz7"] Nov 24 09:05:32 crc kubenswrapper[4944]: I1124 09:05:32.284371 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" path="/var/lib/kubelet/pods/8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd/volumes" Nov 24 09:05:34 crc kubenswrapper[4944]: I1124 09:05:34.513189 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:34 crc kubenswrapper[4944]: I1124 09:05:34.513811 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:34 crc kubenswrapper[4944]: I1124 09:05:34.553836 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:35 crc kubenswrapper[4944]: I1124 09:05:35.518755 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:35 crc kubenswrapper[4944]: I1124 09:05:35.518833 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:35 crc kubenswrapper[4944]: I1124 09:05:35.576752 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:44 crc kubenswrapper[4944]: I1124 09:05:44.560989 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:45 crc kubenswrapper[4944]: I1124 09:05:45.207660 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-66c67cfc8f-mnxqx" Nov 24 09:05:45 crc kubenswrapper[4944]: I1124 09:05:45.644616 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.187080 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx9ww"] Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.188537 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kx9ww" podUID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" containerName="registry-server" containerID="cri-o://cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b" gracePeriod=2 Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.560259 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.673948 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-catalog-content\") pod \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.674088 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-utilities\") pod \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.674131 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqhvv\" (UniqueName: \"kubernetes.io/projected/62372b3b-e2f5-40e6-835c-eb3405c22cfb-kube-api-access-jqhvv\") pod \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\" (UID: \"62372b3b-e2f5-40e6-835c-eb3405c22cfb\") " Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.674900 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-utilities" (OuterVolumeSpecName: "utilities") pod "62372b3b-e2f5-40e6-835c-eb3405c22cfb" (UID: "62372b3b-e2f5-40e6-835c-eb3405c22cfb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.680299 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62372b3b-e2f5-40e6-835c-eb3405c22cfb-kube-api-access-jqhvv" (OuterVolumeSpecName: "kube-api-access-jqhvv") pod "62372b3b-e2f5-40e6-835c-eb3405c22cfb" (UID: "62372b3b-e2f5-40e6-835c-eb3405c22cfb"). InnerVolumeSpecName "kube-api-access-jqhvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.691793 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62372b3b-e2f5-40e6-835c-eb3405c22cfb" (UID: "62372b3b-e2f5-40e6-835c-eb3405c22cfb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.775079 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.775111 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqhvv\" (UniqueName: \"kubernetes.io/projected/62372b3b-e2f5-40e6-835c-eb3405c22cfb-kube-api-access-jqhvv\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:48 crc kubenswrapper[4944]: I1124 09:05:48.775135 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62372b3b-e2f5-40e6-835c-eb3405c22cfb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.056949 4944 generic.go:334] "Generic (PLEG): container finished" podID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" containerID="cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b" exitCode=0 Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.056991 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx9ww" event={"ID":"62372b3b-e2f5-40e6-835c-eb3405c22cfb","Type":"ContainerDied","Data":"cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b"} Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.057017 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kx9ww" event={"ID":"62372b3b-e2f5-40e6-835c-eb3405c22cfb","Type":"ContainerDied","Data":"b146d34ad1e6ee8f445c427c8c3443f91b3319473e2fc84667eb07f1f7122b85"} Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.057035 4944 scope.go:117] "RemoveContainer" containerID="cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b" Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.057419 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kx9ww" Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.077121 4944 scope.go:117] "RemoveContainer" containerID="78ee99ae4ae67ceb1bfc3a11b4ae9e490fce7f040ff8d2fd0699413da6b5186e" Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.090330 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx9ww"] Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.093342 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kx9ww"] Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.110444 4944 scope.go:117] "RemoveContainer" containerID="e3ebcd72da6af2ed9a16a466fba8394ec04c25746bca391860f59a7c0d278551" Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.124559 4944 scope.go:117] "RemoveContainer" containerID="cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b" Nov 24 09:05:49 crc kubenswrapper[4944]: E1124 09:05:49.126380 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b\": container with ID starting with cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b not found: ID does not exist" containerID="cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b" Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.126420 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b"} err="failed to get container status \"cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b\": rpc error: code = NotFound desc = could not find container \"cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b\": container with ID starting with cebc5d65bb7f10a6c33787367fb748e3fd95153e080f5d745bacf6e0cd0b3e6b not found: ID does not exist" Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.126444 4944 scope.go:117] "RemoveContainer" containerID="78ee99ae4ae67ceb1bfc3a11b4ae9e490fce7f040ff8d2fd0699413da6b5186e" Nov 24 09:05:49 crc kubenswrapper[4944]: E1124 09:05:49.126676 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78ee99ae4ae67ceb1bfc3a11b4ae9e490fce7f040ff8d2fd0699413da6b5186e\": container with ID starting with 78ee99ae4ae67ceb1bfc3a11b4ae9e490fce7f040ff8d2fd0699413da6b5186e not found: ID does not exist" containerID="78ee99ae4ae67ceb1bfc3a11b4ae9e490fce7f040ff8d2fd0699413da6b5186e" Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.126705 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78ee99ae4ae67ceb1bfc3a11b4ae9e490fce7f040ff8d2fd0699413da6b5186e"} err="failed to get container status \"78ee99ae4ae67ceb1bfc3a11b4ae9e490fce7f040ff8d2fd0699413da6b5186e\": rpc error: code = NotFound desc = could not find container \"78ee99ae4ae67ceb1bfc3a11b4ae9e490fce7f040ff8d2fd0699413da6b5186e\": container with ID starting with 78ee99ae4ae67ceb1bfc3a11b4ae9e490fce7f040ff8d2fd0699413da6b5186e not found: ID does not exist" Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.126725 4944 scope.go:117] "RemoveContainer" containerID="e3ebcd72da6af2ed9a16a466fba8394ec04c25746bca391860f59a7c0d278551" Nov 24 09:05:49 crc kubenswrapper[4944]: E1124 09:05:49.126939 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3ebcd72da6af2ed9a16a466fba8394ec04c25746bca391860f59a7c0d278551\": container with ID starting with e3ebcd72da6af2ed9a16a466fba8394ec04c25746bca391860f59a7c0d278551 not found: ID does not exist" containerID="e3ebcd72da6af2ed9a16a466fba8394ec04c25746bca391860f59a7c0d278551" Nov 24 09:05:49 crc kubenswrapper[4944]: I1124 09:05:49.126966 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3ebcd72da6af2ed9a16a466fba8394ec04c25746bca391860f59a7c0d278551"} err="failed to get container status \"e3ebcd72da6af2ed9a16a466fba8394ec04c25746bca391860f59a7c0d278551\": rpc error: code = NotFound desc = could not find container \"e3ebcd72da6af2ed9a16a466fba8394ec04c25746bca391860f59a7c0d278551\": container with ID starting with e3ebcd72da6af2ed9a16a466fba8394ec04c25746bca391860f59a7c0d278551 not found: ID does not exist" Nov 24 09:05:50 crc kubenswrapper[4944]: I1124 09:05:50.284558 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" path="/var/lib/kubelet/pods/62372b3b-e2f5-40e6-835c-eb3405c22cfb/volumes" Nov 24 09:05:50 crc kubenswrapper[4944]: I1124 09:05:50.384645 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m2hr9"] Nov 24 09:05:50 crc kubenswrapper[4944]: I1124 09:05:50.384897 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m2hr9" podUID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" containerName="registry-server" containerID="cri-o://5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab" gracePeriod=2 Nov 24 09:05:50 crc kubenswrapper[4944]: I1124 09:05:50.797486 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:50 crc kubenswrapper[4944]: I1124 09:05:50.910313 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-utilities\") pod \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " Nov 24 09:05:50 crc kubenswrapper[4944]: I1124 09:05:50.910438 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-catalog-content\") pod \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " Nov 24 09:05:50 crc kubenswrapper[4944]: I1124 09:05:50.910463 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68jcx\" (UniqueName: \"kubernetes.io/projected/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-kube-api-access-68jcx\") pod \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\" (UID: \"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3\") " Nov 24 09:05:50 crc kubenswrapper[4944]: I1124 09:05:50.911427 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-utilities" (OuterVolumeSpecName: "utilities") pod "a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" (UID: "a5ffe08c-6b4b-41b1-8407-dc51da72ccb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:05:50 crc kubenswrapper[4944]: I1124 09:05:50.916474 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-kube-api-access-68jcx" (OuterVolumeSpecName: "kube-api-access-68jcx") pod "a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" (UID: "a5ffe08c-6b4b-41b1-8407-dc51da72ccb3"). InnerVolumeSpecName "kube-api-access-68jcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:05:50 crc kubenswrapper[4944]: I1124 09:05:50.957842 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" (UID: "a5ffe08c-6b4b-41b1-8407-dc51da72ccb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.012384 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.012412 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68jcx\" (UniqueName: \"kubernetes.io/projected/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-kube-api-access-68jcx\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.012423 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.073178 4944 generic.go:334] "Generic (PLEG): container finished" podID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" containerID="5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab" exitCode=0 Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.073230 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hr9" event={"ID":"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3","Type":"ContainerDied","Data":"5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab"} Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.073480 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m2hr9" event={"ID":"a5ffe08c-6b4b-41b1-8407-dc51da72ccb3","Type":"ContainerDied","Data":"f1e13923cadcdadbbaa8040f4235bc0d599211c81bf6bc7ab32de10d8edb1fb9"} Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.073311 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m2hr9" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.073555 4944 scope.go:117] "RemoveContainer" containerID="5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.099636 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m2hr9"] Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.101932 4944 scope.go:117] "RemoveContainer" containerID="bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.102801 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m2hr9"] Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.143781 4944 scope.go:117] "RemoveContainer" containerID="560c1710dfc67f525235706acaee312ac6365c0bbc72ab09d34378aba4223d16" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.158597 4944 scope.go:117] "RemoveContainer" containerID="5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab" Nov 24 09:05:51 crc kubenswrapper[4944]: E1124 09:05:51.159137 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab\": container with ID starting with 5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab not found: ID does not exist" containerID="5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.159220 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab"} err="failed to get container status \"5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab\": rpc error: code = NotFound desc = could not find container \"5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab\": container with ID starting with 5dfcc69e8bbd544034ddd84f033a3ffd2c2534a404cd13247a00c2ab3b9ba4ab not found: ID does not exist" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.159259 4944 scope.go:117] "RemoveContainer" containerID="bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7" Nov 24 09:05:51 crc kubenswrapper[4944]: E1124 09:05:51.159697 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7\": container with ID starting with bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7 not found: ID does not exist" containerID="bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.159744 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7"} err="failed to get container status \"bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7\": rpc error: code = NotFound desc = could not find container \"bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7\": container with ID starting with bbed4210b37886d36d802ea488d205f6a689cef097b308d1a164292a8651caf7 not found: ID does not exist" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.159762 4944 scope.go:117] "RemoveContainer" containerID="560c1710dfc67f525235706acaee312ac6365c0bbc72ab09d34378aba4223d16" Nov 24 09:05:51 crc kubenswrapper[4944]: E1124 09:05:51.160075 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"560c1710dfc67f525235706acaee312ac6365c0bbc72ab09d34378aba4223d16\": container with ID starting with 560c1710dfc67f525235706acaee312ac6365c0bbc72ab09d34378aba4223d16 not found: ID does not exist" containerID="560c1710dfc67f525235706acaee312ac6365c0bbc72ab09d34378aba4223d16" Nov 24 09:05:51 crc kubenswrapper[4944]: I1124 09:05:51.160164 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"560c1710dfc67f525235706acaee312ac6365c0bbc72ab09d34378aba4223d16"} err="failed to get container status \"560c1710dfc67f525235706acaee312ac6365c0bbc72ab09d34378aba4223d16\": rpc error: code = NotFound desc = could not find container \"560c1710dfc67f525235706acaee312ac6365c0bbc72ab09d34378aba4223d16\": container with ID starting with 560c1710dfc67f525235706acaee312ac6365c0bbc72ab09d34378aba4223d16 not found: ID does not exist" Nov 24 09:05:52 crc kubenswrapper[4944]: I1124 09:05:52.283661 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" path="/var/lib/kubelet/pods/a5ffe08c-6b4b-41b1-8407-dc51da72ccb3/volumes" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997343 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5xsdn"] Nov 24 09:05:53 crc kubenswrapper[4944]: E1124 09:05:53.997594 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" containerName="extract-utilities" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997609 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" containerName="extract-utilities" Nov 24 09:05:53 crc kubenswrapper[4944]: E1124 09:05:53.997623 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" containerName="extract-utilities" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997630 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" containerName="extract-utilities" Nov 24 09:05:53 crc kubenswrapper[4944]: E1124 09:05:53.997640 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" containerName="registry-server" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997648 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" containerName="registry-server" Nov 24 09:05:53 crc kubenswrapper[4944]: E1124 09:05:53.997658 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" containerName="extract-utilities" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997666 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" containerName="extract-utilities" Nov 24 09:05:53 crc kubenswrapper[4944]: E1124 09:05:53.997673 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" containerName="registry-server" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997680 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" containerName="registry-server" Nov 24 09:05:53 crc kubenswrapper[4944]: E1124 09:05:53.997705 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" containerName="extract-content" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997712 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" containerName="extract-content" Nov 24 09:05:53 crc kubenswrapper[4944]: E1124 09:05:53.997724 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" containerName="extract-content" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997731 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" containerName="extract-content" Nov 24 09:05:53 crc kubenswrapper[4944]: E1124 09:05:53.997746 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" containerName="extract-content" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997753 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" containerName="extract-content" Nov 24 09:05:53 crc kubenswrapper[4944]: E1124 09:05:53.997764 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" containerName="registry-server" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997771 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" containerName="registry-server" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997886 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a4aef39-55c4-46d1-9c3f-0f84c4a25fcd" containerName="registry-server" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997901 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5ffe08c-6b4b-41b1-8407-dc51da72ccb3" containerName="registry-server" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.997914 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="62372b3b-e2f5-40e6-835c-eb3405c22cfb" containerName="registry-server" Nov 24 09:05:53 crc kubenswrapper[4944]: I1124 09:05:53.998801 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.011193 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5xsdn"] Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.152993 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-catalog-content\") pod \"community-operators-5xsdn\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.153109 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvhr5\" (UniqueName: \"kubernetes.io/projected/14c58c74-c392-4e88-8a2c-ec3e8704aee0-kube-api-access-lvhr5\") pod \"community-operators-5xsdn\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.153235 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-utilities\") pod \"community-operators-5xsdn\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.254100 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvhr5\" (UniqueName: \"kubernetes.io/projected/14c58c74-c392-4e88-8a2c-ec3e8704aee0-kube-api-access-lvhr5\") pod \"community-operators-5xsdn\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.254168 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-utilities\") pod \"community-operators-5xsdn\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.254225 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-catalog-content\") pod \"community-operators-5xsdn\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.254639 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-utilities\") pod \"community-operators-5xsdn\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.254718 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-catalog-content\") pod \"community-operators-5xsdn\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.281234 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvhr5\" (UniqueName: \"kubernetes.io/projected/14c58c74-c392-4e88-8a2c-ec3e8704aee0-kube-api-access-lvhr5\") pod \"community-operators-5xsdn\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.314641 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:05:54 crc kubenswrapper[4944]: I1124 09:05:54.840273 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5xsdn"] Nov 24 09:05:54 crc kubenswrapper[4944]: W1124 09:05:54.849189 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14c58c74_c392_4e88_8a2c_ec3e8704aee0.slice/crio-949585433a1bfe8121f303e44fa3a6d5791cec339febd3a3d51cd8df853bd0c3 WatchSource:0}: Error finding container 949585433a1bfe8121f303e44fa3a6d5791cec339febd3a3d51cd8df853bd0c3: Status 404 returned error can't find the container with id 949585433a1bfe8121f303e44fa3a6d5791cec339febd3a3d51cd8df853bd0c3 Nov 24 09:05:55 crc kubenswrapper[4944]: I1124 09:05:55.108922 4944 generic.go:334] "Generic (PLEG): container finished" podID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" containerID="1e2e97a93d386a6f540412e58de1eb917b3056027afd573b9a6409c9bad247c4" exitCode=0 Nov 24 09:05:55 crc kubenswrapper[4944]: I1124 09:05:55.108968 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xsdn" event={"ID":"14c58c74-c392-4e88-8a2c-ec3e8704aee0","Type":"ContainerDied","Data":"1e2e97a93d386a6f540412e58de1eb917b3056027afd573b9a6409c9bad247c4"} Nov 24 09:05:55 crc kubenswrapper[4944]: I1124 09:05:55.109264 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xsdn" event={"ID":"14c58c74-c392-4e88-8a2c-ec3e8704aee0","Type":"ContainerStarted","Data":"949585433a1bfe8121f303e44fa3a6d5791cec339febd3a3d51cd8df853bd0c3"} Nov 24 09:05:56 crc kubenswrapper[4944]: I1124 09:05:56.117440 4944 generic.go:334] "Generic (PLEG): container finished" podID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" containerID="d5615176e70a88d2fcc7a2aaaca41acd4e262e949d0f39f90fc6988d6b199363" exitCode=0 Nov 24 09:05:56 crc kubenswrapper[4944]: I1124 09:05:56.117476 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xsdn" event={"ID":"14c58c74-c392-4e88-8a2c-ec3e8704aee0","Type":"ContainerDied","Data":"d5615176e70a88d2fcc7a2aaaca41acd4e262e949d0f39f90fc6988d6b199363"} Nov 24 09:05:57 crc kubenswrapper[4944]: I1124 09:05:57.125456 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xsdn" event={"ID":"14c58c74-c392-4e88-8a2c-ec3e8704aee0","Type":"ContainerStarted","Data":"db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0"} Nov 24 09:05:57 crc kubenswrapper[4944]: I1124 09:05:57.143834 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5xsdn" podStartSLOduration=2.743340291 podStartE2EDuration="4.143799509s" podCreationTimestamp="2025-11-24 09:05:53 +0000 UTC" firstStartedPulling="2025-11-24 09:05:55.110513564 +0000 UTC m=+815.644954026" lastFinishedPulling="2025-11-24 09:05:56.510972782 +0000 UTC m=+817.045413244" observedRunningTime="2025-11-24 09:05:57.14192819 +0000 UTC m=+817.676368672" watchObservedRunningTime="2025-11-24 09:05:57.143799509 +0000 UTC m=+817.678239971" Nov 24 09:06:04 crc kubenswrapper[4944]: I1124 09:06:04.315442 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:06:04 crc kubenswrapper[4944]: I1124 09:06:04.317238 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:06:04 crc kubenswrapper[4944]: I1124 09:06:04.353898 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:06:04 crc kubenswrapper[4944]: I1124 09:06:04.679073 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6c8c4f86d5-h8tbj" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.227354 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.444576 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-jbcpf"] Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.447654 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.448928 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-brg9g"] Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.449460 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.450375 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.450781 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.450776 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-ds548" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.450895 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.465105 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-brg9g"] Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.537977 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-metrics\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.538116 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-frr-sockets\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.538203 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23ee832e-3f27-47bc-b8cb-67e0df2f5b74-cert\") pod \"frr-k8s-webhook-server-6998585d5-brg9g\" (UID: \"23ee832e-3f27-47bc-b8cb-67e0df2f5b74\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.538268 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-reloader\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.538358 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-metrics-certs\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.538398 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdcr9\" (UniqueName: \"kubernetes.io/projected/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-kube-api-access-kdcr9\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.538417 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-frr-conf\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.538434 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-frr-startup\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.538459 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds9wp\" (UniqueName: \"kubernetes.io/projected/23ee832e-3f27-47bc-b8cb-67e0df2f5b74-kube-api-access-ds9wp\") pod \"frr-k8s-webhook-server-6998585d5-brg9g\" (UID: \"23ee832e-3f27-47bc-b8cb-67e0df2f5b74\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.566407 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-nl6qg"] Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.567357 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nl6qg" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.573085 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dzn7v" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.573245 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.574087 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.582081 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-hfh7t"] Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.583266 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.583539 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.585299 4944 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.591383 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-hfh7t"] Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.639335 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdcr9\" (UniqueName: \"kubernetes.io/projected/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-kube-api-access-kdcr9\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.639396 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-frr-conf\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.639429 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-frr-startup\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.639634 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds9wp\" (UniqueName: \"kubernetes.io/projected/23ee832e-3f27-47bc-b8cb-67e0df2f5b74-kube-api-access-ds9wp\") pod \"frr-k8s-webhook-server-6998585d5-brg9g\" (UID: \"23ee832e-3f27-47bc-b8cb-67e0df2f5b74\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.639735 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-metrics\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.639838 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-frr-sockets\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.639925 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-reloader\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.639958 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23ee832e-3f27-47bc-b8cb-67e0df2f5b74-cert\") pod \"frr-k8s-webhook-server-6998585d5-brg9g\" (UID: \"23ee832e-3f27-47bc-b8cb-67e0df2f5b74\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.640006 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-metrics-certs\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: E1124 09:06:05.640196 4944 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Nov 24 09:06:05 crc kubenswrapper[4944]: E1124 09:06:05.640317 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/23ee832e-3f27-47bc-b8cb-67e0df2f5b74-cert podName:23ee832e-3f27-47bc-b8cb-67e0df2f5b74 nodeName:}" failed. No retries permitted until 2025-11-24 09:06:06.140287365 +0000 UTC m=+826.674727827 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/23ee832e-3f27-47bc-b8cb-67e0df2f5b74-cert") pod "frr-k8s-webhook-server-6998585d5-brg9g" (UID: "23ee832e-3f27-47bc-b8cb-67e0df2f5b74") : secret "frr-k8s-webhook-server-cert" not found Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.640820 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-frr-startup\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.640842 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-frr-conf\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.640862 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-frr-sockets\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.641022 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-metrics\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.641159 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-reloader\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.658162 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds9wp\" (UniqueName: \"kubernetes.io/projected/23ee832e-3f27-47bc-b8cb-67e0df2f5b74-kube-api-access-ds9wp\") pod \"frr-k8s-webhook-server-6998585d5-brg9g\" (UID: \"23ee832e-3f27-47bc-b8cb-67e0df2f5b74\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.660677 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdcr9\" (UniqueName: \"kubernetes.io/projected/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-kube-api-access-kdcr9\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.665582 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61a2baaa-dc80-42dc-b458-0e9ffb5d0e77-metrics-certs\") pod \"frr-k8s-jbcpf\" (UID: \"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77\") " pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.740710 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr7v5\" (UniqueName: \"kubernetes.io/projected/a2469128-b8f5-46f4-80a4-fe4a25aed791-kube-api-access-tr7v5\") pod \"controller-6c7b4b5f48-hfh7t\" (UID: \"a2469128-b8f5-46f4-80a4-fe4a25aed791\") " pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.740773 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96hks\" (UniqueName: \"kubernetes.io/projected/169a7266-9eeb-4984-86d9-6b49c647a622-kube-api-access-96hks\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.740882 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2469128-b8f5-46f4-80a4-fe4a25aed791-cert\") pod \"controller-6c7b4b5f48-hfh7t\" (UID: \"a2469128-b8f5-46f4-80a4-fe4a25aed791\") " pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.740970 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a2469128-b8f5-46f4-80a4-fe4a25aed791-metrics-certs\") pod \"controller-6c7b4b5f48-hfh7t\" (UID: \"a2469128-b8f5-46f4-80a4-fe4a25aed791\") " pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.740986 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-metrics-certs\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.741085 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/169a7266-9eeb-4984-86d9-6b49c647a622-metallb-excludel2\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.741140 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-memberlist\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.794000 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.842475 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96hks\" (UniqueName: \"kubernetes.io/projected/169a7266-9eeb-4984-86d9-6b49c647a622-kube-api-access-96hks\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.842551 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2469128-b8f5-46f4-80a4-fe4a25aed791-cert\") pod \"controller-6c7b4b5f48-hfh7t\" (UID: \"a2469128-b8f5-46f4-80a4-fe4a25aed791\") " pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.842608 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-metrics-certs\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.842626 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a2469128-b8f5-46f4-80a4-fe4a25aed791-metrics-certs\") pod \"controller-6c7b4b5f48-hfh7t\" (UID: \"a2469128-b8f5-46f4-80a4-fe4a25aed791\") " pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.842658 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/169a7266-9eeb-4984-86d9-6b49c647a622-metallb-excludel2\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.842680 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-memberlist\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.842722 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr7v5\" (UniqueName: \"kubernetes.io/projected/a2469128-b8f5-46f4-80a4-fe4a25aed791-kube-api-access-tr7v5\") pod \"controller-6c7b4b5f48-hfh7t\" (UID: \"a2469128-b8f5-46f4-80a4-fe4a25aed791\") " pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:05 crc kubenswrapper[4944]: E1124 09:06:05.846217 4944 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Nov 24 09:06:05 crc kubenswrapper[4944]: E1124 09:06:05.846323 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-metrics-certs podName:169a7266-9eeb-4984-86d9-6b49c647a622 nodeName:}" failed. No retries permitted until 2025-11-24 09:06:06.346298194 +0000 UTC m=+826.880738656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-metrics-certs") pod "speaker-nl6qg" (UID: "169a7266-9eeb-4984-86d9-6b49c647a622") : secret "speaker-certs-secret" not found Nov 24 09:06:05 crc kubenswrapper[4944]: E1124 09:06:05.846350 4944 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 09:06:05 crc kubenswrapper[4944]: E1124 09:06:05.846442 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-memberlist podName:169a7266-9eeb-4984-86d9-6b49c647a622 nodeName:}" failed. No retries permitted until 2025-11-24 09:06:06.346419878 +0000 UTC m=+826.880860400 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-memberlist") pod "speaker-nl6qg" (UID: "169a7266-9eeb-4984-86d9-6b49c647a622") : secret "metallb-memberlist" not found Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.846917 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/169a7266-9eeb-4984-86d9-6b49c647a622-metallb-excludel2\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.851156 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2469128-b8f5-46f4-80a4-fe4a25aed791-cert\") pod \"controller-6c7b4b5f48-hfh7t\" (UID: \"a2469128-b8f5-46f4-80a4-fe4a25aed791\") " pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.853148 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a2469128-b8f5-46f4-80a4-fe4a25aed791-metrics-certs\") pod \"controller-6c7b4b5f48-hfh7t\" (UID: \"a2469128-b8f5-46f4-80a4-fe4a25aed791\") " pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.874610 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96hks\" (UniqueName: \"kubernetes.io/projected/169a7266-9eeb-4984-86d9-6b49c647a622-kube-api-access-96hks\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.876289 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr7v5\" (UniqueName: \"kubernetes.io/projected/a2469128-b8f5-46f4-80a4-fe4a25aed791-kube-api-access-tr7v5\") pod \"controller-6c7b4b5f48-hfh7t\" (UID: \"a2469128-b8f5-46f4-80a4-fe4a25aed791\") " pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:05 crc kubenswrapper[4944]: I1124 09:06:05.897982 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:06 crc kubenswrapper[4944]: I1124 09:06:06.151035 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23ee832e-3f27-47bc-b8cb-67e0df2f5b74-cert\") pod \"frr-k8s-webhook-server-6998585d5-brg9g\" (UID: \"23ee832e-3f27-47bc-b8cb-67e0df2f5b74\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" Nov 24 09:06:06 crc kubenswrapper[4944]: I1124 09:06:06.157298 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/23ee832e-3f27-47bc-b8cb-67e0df2f5b74-cert\") pod \"frr-k8s-webhook-server-6998585d5-brg9g\" (UID: \"23ee832e-3f27-47bc-b8cb-67e0df2f5b74\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" Nov 24 09:06:06 crc kubenswrapper[4944]: I1124 09:06:06.195613 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbcpf" event={"ID":"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77","Type":"ContainerStarted","Data":"1b92c8924133af866a025843da4e9d8a77e9c1b33cc20c36382c5494e845cdb9"} Nov 24 09:06:06 crc kubenswrapper[4944]: I1124 09:06:06.353247 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-metrics-certs\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:06 crc kubenswrapper[4944]: I1124 09:06:06.353553 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-memberlist\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:06 crc kubenswrapper[4944]: E1124 09:06:06.354561 4944 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 09:06:06 crc kubenswrapper[4944]: E1124 09:06:06.354936 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-memberlist podName:169a7266-9eeb-4984-86d9-6b49c647a622 nodeName:}" failed. No retries permitted until 2025-11-24 09:06:07.354898987 +0000 UTC m=+827.889339449 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-memberlist") pod "speaker-nl6qg" (UID: "169a7266-9eeb-4984-86d9-6b49c647a622") : secret "metallb-memberlist" not found Nov 24 09:06:06 crc kubenswrapper[4944]: I1124 09:06:06.358375 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-metrics-certs\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:06 crc kubenswrapper[4944]: I1124 09:06:06.366220 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-hfh7t"] Nov 24 09:06:06 crc kubenswrapper[4944]: W1124 09:06:06.374881 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2469128_b8f5_46f4_80a4_fe4a25aed791.slice/crio-cfdcb442b0ff35ac13791e42024c25fdc13413e6f28dbbe670ebdb7871d46af1 WatchSource:0}: Error finding container cfdcb442b0ff35ac13791e42024c25fdc13413e6f28dbbe670ebdb7871d46af1: Status 404 returned error can't find the container with id cfdcb442b0ff35ac13791e42024c25fdc13413e6f28dbbe670ebdb7871d46af1 Nov 24 09:06:06 crc kubenswrapper[4944]: I1124 09:06:06.392225 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" Nov 24 09:06:06 crc kubenswrapper[4944]: I1124 09:06:06.586692 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5xsdn"] Nov 24 09:06:06 crc kubenswrapper[4944]: I1124 09:06:06.783908 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-brg9g"] Nov 24 09:06:06 crc kubenswrapper[4944]: W1124 09:06:06.790674 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23ee832e_3f27_47bc_b8cb_67e0df2f5b74.slice/crio-8bdc50f58a9150954ba19894ffe31319628873dac83cbaf678c569e454d6ca66 WatchSource:0}: Error finding container 8bdc50f58a9150954ba19894ffe31319628873dac83cbaf678c569e454d6ca66: Status 404 returned error can't find the container with id 8bdc50f58a9150954ba19894ffe31319628873dac83cbaf678c569e454d6ca66 Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.205638 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-hfh7t" event={"ID":"a2469128-b8f5-46f4-80a4-fe4a25aed791","Type":"ContainerStarted","Data":"4c7207e0657e495d4a4c81b146dfea44a3398f1a0cc12f454d861c1a341e1677"} Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.205690 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-hfh7t" event={"ID":"a2469128-b8f5-46f4-80a4-fe4a25aed791","Type":"ContainerStarted","Data":"584ca863a29fab6e8471ddca93bca66441f55630e1384b3ee2c089c11aa896db"} Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.205704 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-hfh7t" event={"ID":"a2469128-b8f5-46f4-80a4-fe4a25aed791","Type":"ContainerStarted","Data":"cfdcb442b0ff35ac13791e42024c25fdc13413e6f28dbbe670ebdb7871d46af1"} Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.205741 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.208359 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" event={"ID":"23ee832e-3f27-47bc-b8cb-67e0df2f5b74","Type":"ContainerStarted","Data":"8bdc50f58a9150954ba19894ffe31319628873dac83cbaf678c569e454d6ca66"} Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.208473 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5xsdn" podUID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" containerName="registry-server" containerID="cri-o://db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0" gracePeriod=2 Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.227390 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-hfh7t" podStartSLOduration=2.227367314 podStartE2EDuration="2.227367314s" podCreationTimestamp="2025-11-24 09:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:06:07.222563511 +0000 UTC m=+827.757003993" watchObservedRunningTime="2025-11-24 09:06:07.227367314 +0000 UTC m=+827.761807776" Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.367037 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-memberlist\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.396438 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/169a7266-9eeb-4984-86d9-6b49c647a622-memberlist\") pod \"speaker-nl6qg\" (UID: \"169a7266-9eeb-4984-86d9-6b49c647a622\") " pod="metallb-system/speaker-nl6qg" Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.684961 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nl6qg" Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.771423 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.876736 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvhr5\" (UniqueName: \"kubernetes.io/projected/14c58c74-c392-4e88-8a2c-ec3e8704aee0-kube-api-access-lvhr5\") pod \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.876898 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-utilities\") pod \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.876968 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-catalog-content\") pod \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\" (UID: \"14c58c74-c392-4e88-8a2c-ec3e8704aee0\") " Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.877919 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-utilities" (OuterVolumeSpecName: "utilities") pod "14c58c74-c392-4e88-8a2c-ec3e8704aee0" (UID: "14c58c74-c392-4e88-8a2c-ec3e8704aee0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.881032 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14c58c74-c392-4e88-8a2c-ec3e8704aee0-kube-api-access-lvhr5" (OuterVolumeSpecName: "kube-api-access-lvhr5") pod "14c58c74-c392-4e88-8a2c-ec3e8704aee0" (UID: "14c58c74-c392-4e88-8a2c-ec3e8704aee0"). InnerVolumeSpecName "kube-api-access-lvhr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.943141 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14c58c74-c392-4e88-8a2c-ec3e8704aee0" (UID: "14c58c74-c392-4e88-8a2c-ec3e8704aee0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.981106 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.981162 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvhr5\" (UniqueName: \"kubernetes.io/projected/14c58c74-c392-4e88-8a2c-ec3e8704aee0-kube-api-access-lvhr5\") on node \"crc\" DevicePath \"\"" Nov 24 09:06:07 crc kubenswrapper[4944]: I1124 09:06:07.981177 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14c58c74-c392-4e88-8a2c-ec3e8704aee0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.229412 4944 generic.go:334] "Generic (PLEG): container finished" podID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" containerID="db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0" exitCode=0 Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.229516 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xsdn" Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.230136 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xsdn" event={"ID":"14c58c74-c392-4e88-8a2c-ec3e8704aee0","Type":"ContainerDied","Data":"db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0"} Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.230177 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xsdn" event={"ID":"14c58c74-c392-4e88-8a2c-ec3e8704aee0","Type":"ContainerDied","Data":"949585433a1bfe8121f303e44fa3a6d5791cec339febd3a3d51cd8df853bd0c3"} Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.230207 4944 scope.go:117] "RemoveContainer" containerID="db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0" Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.242698 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nl6qg" event={"ID":"169a7266-9eeb-4984-86d9-6b49c647a622","Type":"ContainerStarted","Data":"b23612318a672fd8eab245c499ea1480eecd4eeb07264ac6b77979f4fdcc155f"} Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.273684 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5xsdn"] Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.274059 4944 scope.go:117] "RemoveContainer" containerID="d5615176e70a88d2fcc7a2aaaca41acd4e262e949d0f39f90fc6988d6b199363" Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.286460 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5xsdn"] Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.313451 4944 scope.go:117] "RemoveContainer" containerID="1e2e97a93d386a6f540412e58de1eb917b3056027afd573b9a6409c9bad247c4" Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.332998 4944 scope.go:117] "RemoveContainer" containerID="db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0" Nov 24 09:06:08 crc kubenswrapper[4944]: E1124 09:06:08.333465 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0\": container with ID starting with db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0 not found: ID does not exist" containerID="db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0" Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.333525 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0"} err="failed to get container status \"db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0\": rpc error: code = NotFound desc = could not find container \"db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0\": container with ID starting with db1eda1cdc465d9ec00f3df43050512b1a0842ce9ecfb3140345a1ac36d147c0 not found: ID does not exist" Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.333565 4944 scope.go:117] "RemoveContainer" containerID="d5615176e70a88d2fcc7a2aaaca41acd4e262e949d0f39f90fc6988d6b199363" Nov 24 09:06:08 crc kubenswrapper[4944]: E1124 09:06:08.333952 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5615176e70a88d2fcc7a2aaaca41acd4e262e949d0f39f90fc6988d6b199363\": container with ID starting with d5615176e70a88d2fcc7a2aaaca41acd4e262e949d0f39f90fc6988d6b199363 not found: ID does not exist" containerID="d5615176e70a88d2fcc7a2aaaca41acd4e262e949d0f39f90fc6988d6b199363" Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.334079 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5615176e70a88d2fcc7a2aaaca41acd4e262e949d0f39f90fc6988d6b199363"} err="failed to get container status \"d5615176e70a88d2fcc7a2aaaca41acd4e262e949d0f39f90fc6988d6b199363\": rpc error: code = NotFound desc = could not find container \"d5615176e70a88d2fcc7a2aaaca41acd4e262e949d0f39f90fc6988d6b199363\": container with ID starting with d5615176e70a88d2fcc7a2aaaca41acd4e262e949d0f39f90fc6988d6b199363 not found: ID does not exist" Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.334121 4944 scope.go:117] "RemoveContainer" containerID="1e2e97a93d386a6f540412e58de1eb917b3056027afd573b9a6409c9bad247c4" Nov 24 09:06:08 crc kubenswrapper[4944]: E1124 09:06:08.334516 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e2e97a93d386a6f540412e58de1eb917b3056027afd573b9a6409c9bad247c4\": container with ID starting with 1e2e97a93d386a6f540412e58de1eb917b3056027afd573b9a6409c9bad247c4 not found: ID does not exist" containerID="1e2e97a93d386a6f540412e58de1eb917b3056027afd573b9a6409c9bad247c4" Nov 24 09:06:08 crc kubenswrapper[4944]: I1124 09:06:08.334550 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e2e97a93d386a6f540412e58de1eb917b3056027afd573b9a6409c9bad247c4"} err="failed to get container status \"1e2e97a93d386a6f540412e58de1eb917b3056027afd573b9a6409c9bad247c4\": rpc error: code = NotFound desc = could not find container \"1e2e97a93d386a6f540412e58de1eb917b3056027afd573b9a6409c9bad247c4\": container with ID starting with 1e2e97a93d386a6f540412e58de1eb917b3056027afd573b9a6409c9bad247c4 not found: ID does not exist" Nov 24 09:06:09 crc kubenswrapper[4944]: I1124 09:06:09.289832 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nl6qg" event={"ID":"169a7266-9eeb-4984-86d9-6b49c647a622","Type":"ContainerStarted","Data":"baee0b24f9058b50d58879a97642271e8371b12763ef005a5c3d2b66ef81f42d"} Nov 24 09:06:09 crc kubenswrapper[4944]: I1124 09:06:09.290124 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nl6qg" event={"ID":"169a7266-9eeb-4984-86d9-6b49c647a622","Type":"ContainerStarted","Data":"4a5d34577716ac4e7f16b9e60a5fe58947c51f4d77752762b6f607e51e36e545"} Nov 24 09:06:09 crc kubenswrapper[4944]: I1124 09:06:09.290856 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-nl6qg" Nov 24 09:06:10 crc kubenswrapper[4944]: I1124 09:06:10.288168 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" path="/var/lib/kubelet/pods/14c58c74-c392-4e88-8a2c-ec3e8704aee0/volumes" Nov 24 09:06:10 crc kubenswrapper[4944]: I1124 09:06:10.298871 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-nl6qg" podStartSLOduration=5.298847932 podStartE2EDuration="5.298847932s" podCreationTimestamp="2025-11-24 09:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:06:09.327636611 +0000 UTC m=+829.862077083" watchObservedRunningTime="2025-11-24 09:06:10.298847932 +0000 UTC m=+830.833288394" Nov 24 09:06:14 crc kubenswrapper[4944]: I1124 09:06:14.332732 4944 generic.go:334] "Generic (PLEG): container finished" podID="61a2baaa-dc80-42dc-b458-0e9ffb5d0e77" containerID="3b49b5a9f8576c73120961e835012fc187a0b42cb5a52eefe198f1fa0145ef55" exitCode=0 Nov 24 09:06:14 crc kubenswrapper[4944]: I1124 09:06:14.332782 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbcpf" event={"ID":"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77","Type":"ContainerDied","Data":"3b49b5a9f8576c73120961e835012fc187a0b42cb5a52eefe198f1fa0145ef55"} Nov 24 09:06:14 crc kubenswrapper[4944]: I1124 09:06:14.337753 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" event={"ID":"23ee832e-3f27-47bc-b8cb-67e0df2f5b74","Type":"ContainerStarted","Data":"55b9486f1d9408c2b070d4be6a860a37dd65c7f7b5f1847b947e65f15153c1bf"} Nov 24 09:06:14 crc kubenswrapper[4944]: I1124 09:06:14.369937 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" podStartSLOduration=2.4595623460000002 podStartE2EDuration="9.369911884s" podCreationTimestamp="2025-11-24 09:06:05 +0000 UTC" firstStartedPulling="2025-11-24 09:06:06.793659786 +0000 UTC m=+827.328100248" lastFinishedPulling="2025-11-24 09:06:13.704009324 +0000 UTC m=+834.238449786" observedRunningTime="2025-11-24 09:06:14.368477908 +0000 UTC m=+834.902918370" watchObservedRunningTime="2025-11-24 09:06:14.369911884 +0000 UTC m=+834.904352356" Nov 24 09:06:15 crc kubenswrapper[4944]: I1124 09:06:15.344211 4944 generic.go:334] "Generic (PLEG): container finished" podID="61a2baaa-dc80-42dc-b458-0e9ffb5d0e77" containerID="0ae684171dca559afb5139d3575cd4eed9e566a68d3ec6a1601ea11a2936eba8" exitCode=0 Nov 24 09:06:15 crc kubenswrapper[4944]: I1124 09:06:15.344288 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbcpf" event={"ID":"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77","Type":"ContainerDied","Data":"0ae684171dca559afb5139d3575cd4eed9e566a68d3ec6a1601ea11a2936eba8"} Nov 24 09:06:15 crc kubenswrapper[4944]: I1124 09:06:15.344670 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" Nov 24 09:06:16 crc kubenswrapper[4944]: I1124 09:06:16.356439 4944 generic.go:334] "Generic (PLEG): container finished" podID="61a2baaa-dc80-42dc-b458-0e9ffb5d0e77" containerID="19988eb553c499ed66daa24295f2bb818aab58a940d897f36c4d61fa0e30af79" exitCode=0 Nov 24 09:06:16 crc kubenswrapper[4944]: I1124 09:06:16.356583 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbcpf" event={"ID":"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77","Type":"ContainerDied","Data":"19988eb553c499ed66daa24295f2bb818aab58a940d897f36c4d61fa0e30af79"} Nov 24 09:06:17 crc kubenswrapper[4944]: I1124 09:06:17.374466 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbcpf" event={"ID":"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77","Type":"ContainerStarted","Data":"e65ee600cf2ec6deab81fb6eca004eca05dd8644891eaae117d3fd4b82062981"} Nov 24 09:06:17 crc kubenswrapper[4944]: I1124 09:06:17.374942 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbcpf" event={"ID":"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77","Type":"ContainerStarted","Data":"a2ca89b8ba7133687056d1a1d768dc4599580b5b25419f1cced74bceb2072ef5"} Nov 24 09:06:17 crc kubenswrapper[4944]: I1124 09:06:17.374964 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbcpf" event={"ID":"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77","Type":"ContainerStarted","Data":"edbb59d83157cc45a3d8add43cb734ebc426aeaa5da7f4c68cc69acf9fd41f04"} Nov 24 09:06:17 crc kubenswrapper[4944]: I1124 09:06:17.374979 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbcpf" event={"ID":"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77","Type":"ContainerStarted","Data":"df43fbb5f29f2e763a0480d6609e8716d0eee0d6335895c5c342295b9b6c9bbf"} Nov 24 09:06:17 crc kubenswrapper[4944]: I1124 09:06:17.374991 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbcpf" event={"ID":"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77","Type":"ContainerStarted","Data":"e7060cfb1ccdf924e30b6e364de388c3165a4d1b66a58f854b982629773c1d37"} Nov 24 09:06:18 crc kubenswrapper[4944]: I1124 09:06:18.384566 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbcpf" event={"ID":"61a2baaa-dc80-42dc-b458-0e9ffb5d0e77","Type":"ContainerStarted","Data":"da4a5bbdb5b446f73484bbe3e0172db1c8ca7eaab5e31a336fe5386352391624"} Nov 24 09:06:18 crc kubenswrapper[4944]: I1124 09:06:18.384733 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:18 crc kubenswrapper[4944]: I1124 09:06:18.408187 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-jbcpf" podStartSLOduration=5.730919227 podStartE2EDuration="13.408169961s" podCreationTimestamp="2025-11-24 09:06:05 +0000 UTC" firstStartedPulling="2025-11-24 09:06:06.004245023 +0000 UTC m=+826.538685475" lastFinishedPulling="2025-11-24 09:06:13.681495747 +0000 UTC m=+834.215936209" observedRunningTime="2025-11-24 09:06:18.407749748 +0000 UTC m=+838.942190220" watchObservedRunningTime="2025-11-24 09:06:18.408169961 +0000 UTC m=+838.942610423" Nov 24 09:06:20 crc kubenswrapper[4944]: I1124 09:06:20.795928 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:20 crc kubenswrapper[4944]: I1124 09:06:20.832424 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:25 crc kubenswrapper[4944]: I1124 09:06:25.901680 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-hfh7t" Nov 24 09:06:26 crc kubenswrapper[4944]: I1124 09:06:26.402425 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-brg9g" Nov 24 09:06:27 crc kubenswrapper[4944]: I1124 09:06:27.689824 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-nl6qg" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.427582 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46"] Nov 24 09:06:29 crc kubenswrapper[4944]: E1124 09:06:29.428138 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" containerName="extract-content" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.428153 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" containerName="extract-content" Nov 24 09:06:29 crc kubenswrapper[4944]: E1124 09:06:29.428175 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" containerName="extract-utilities" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.428181 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" containerName="extract-utilities" Nov 24 09:06:29 crc kubenswrapper[4944]: E1124 09:06:29.428190 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" containerName="registry-server" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.428197 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" containerName="registry-server" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.428299 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="14c58c74-c392-4e88-8a2c-ec3e8704aee0" containerName="registry-server" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.429178 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.430802 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.437384 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46"] Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.584846 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.584905 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.584945 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nwnr\" (UniqueName: \"kubernetes.io/projected/a4e4656e-3009-4e6f-a478-c8cbefc40fba-kube-api-access-7nwnr\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.686683 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nwnr\" (UniqueName: \"kubernetes.io/projected/a4e4656e-3009-4e6f-a478-c8cbefc40fba-kube-api-access-7nwnr\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.686783 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.686805 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.687229 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.687316 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.706800 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nwnr\" (UniqueName: \"kubernetes.io/projected/a4e4656e-3009-4e6f-a478-c8cbefc40fba-kube-api-access-7nwnr\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:29 crc kubenswrapper[4944]: I1124 09:06:29.758944 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:30 crc kubenswrapper[4944]: I1124 09:06:30.214756 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46"] Nov 24 09:06:30 crc kubenswrapper[4944]: W1124 09:06:30.226231 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4e4656e_3009_4e6f_a478_c8cbefc40fba.slice/crio-f6c39c0e22d53944b7fe4d7dc31ad226e0deb1c01152852fd9324cf576914ca3 WatchSource:0}: Error finding container f6c39c0e22d53944b7fe4d7dc31ad226e0deb1c01152852fd9324cf576914ca3: Status 404 returned error can't find the container with id f6c39c0e22d53944b7fe4d7dc31ad226e0deb1c01152852fd9324cf576914ca3 Nov 24 09:06:30 crc kubenswrapper[4944]: I1124 09:06:30.454245 4944 generic.go:334] "Generic (PLEG): container finished" podID="a4e4656e-3009-4e6f-a478-c8cbefc40fba" containerID="0cd14f5f70f37aa71009c100073986755cf9eb985739bc84d35e5fbf1fba00a6" exitCode=0 Nov 24 09:06:30 crc kubenswrapper[4944]: I1124 09:06:30.454337 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" event={"ID":"a4e4656e-3009-4e6f-a478-c8cbefc40fba","Type":"ContainerDied","Data":"0cd14f5f70f37aa71009c100073986755cf9eb985739bc84d35e5fbf1fba00a6"} Nov 24 09:06:30 crc kubenswrapper[4944]: I1124 09:06:30.454638 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" event={"ID":"a4e4656e-3009-4e6f-a478-c8cbefc40fba","Type":"ContainerStarted","Data":"f6c39c0e22d53944b7fe4d7dc31ad226e0deb1c01152852fd9324cf576914ca3"} Nov 24 09:06:34 crc kubenswrapper[4944]: I1124 09:06:34.478654 4944 generic.go:334] "Generic (PLEG): container finished" podID="a4e4656e-3009-4e6f-a478-c8cbefc40fba" containerID="4c9e134ed48fe7cb8e750e0dd887b27885580ecd695ca34c96b054ff0b63c2a8" exitCode=0 Nov 24 09:06:34 crc kubenswrapper[4944]: I1124 09:06:34.478801 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" event={"ID":"a4e4656e-3009-4e6f-a478-c8cbefc40fba","Type":"ContainerDied","Data":"4c9e134ed48fe7cb8e750e0dd887b27885580ecd695ca34c96b054ff0b63c2a8"} Nov 24 09:06:35 crc kubenswrapper[4944]: I1124 09:06:35.488105 4944 generic.go:334] "Generic (PLEG): container finished" podID="a4e4656e-3009-4e6f-a478-c8cbefc40fba" containerID="0d7e7541bcb406bdc2e9df18801b25d8da1fb34fa1719910849ba6ddd858a8eb" exitCode=0 Nov 24 09:06:35 crc kubenswrapper[4944]: I1124 09:06:35.488209 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" event={"ID":"a4e4656e-3009-4e6f-a478-c8cbefc40fba","Type":"ContainerDied","Data":"0d7e7541bcb406bdc2e9df18801b25d8da1fb34fa1719910849ba6ddd858a8eb"} Nov 24 09:06:35 crc kubenswrapper[4944]: I1124 09:06:35.800241 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-jbcpf" Nov 24 09:06:36 crc kubenswrapper[4944]: I1124 09:06:36.751827 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:36 crc kubenswrapper[4944]: I1124 09:06:36.904885 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-bundle\") pod \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " Nov 24 09:06:36 crc kubenswrapper[4944]: I1124 09:06:36.905883 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-bundle" (OuterVolumeSpecName: "bundle") pod "a4e4656e-3009-4e6f-a478-c8cbefc40fba" (UID: "a4e4656e-3009-4e6f-a478-c8cbefc40fba"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:06:36 crc kubenswrapper[4944]: I1124 09:06:36.906007 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-util\") pod \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " Nov 24 09:06:36 crc kubenswrapper[4944]: I1124 09:06:36.906145 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nwnr\" (UniqueName: \"kubernetes.io/projected/a4e4656e-3009-4e6f-a478-c8cbefc40fba-kube-api-access-7nwnr\") pod \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\" (UID: \"a4e4656e-3009-4e6f-a478-c8cbefc40fba\") " Nov 24 09:06:36 crc kubenswrapper[4944]: I1124 09:06:36.907222 4944 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:06:36 crc kubenswrapper[4944]: I1124 09:06:36.911451 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4e4656e-3009-4e6f-a478-c8cbefc40fba-kube-api-access-7nwnr" (OuterVolumeSpecName: "kube-api-access-7nwnr") pod "a4e4656e-3009-4e6f-a478-c8cbefc40fba" (UID: "a4e4656e-3009-4e6f-a478-c8cbefc40fba"). InnerVolumeSpecName "kube-api-access-7nwnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:06:36 crc kubenswrapper[4944]: I1124 09:06:36.925723 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-util" (OuterVolumeSpecName: "util") pod "a4e4656e-3009-4e6f-a478-c8cbefc40fba" (UID: "a4e4656e-3009-4e6f-a478-c8cbefc40fba"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:06:37 crc kubenswrapper[4944]: I1124 09:06:37.009380 4944 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4e4656e-3009-4e6f-a478-c8cbefc40fba-util\") on node \"crc\" DevicePath \"\"" Nov 24 09:06:37 crc kubenswrapper[4944]: I1124 09:06:37.009460 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nwnr\" (UniqueName: \"kubernetes.io/projected/a4e4656e-3009-4e6f-a478-c8cbefc40fba-kube-api-access-7nwnr\") on node \"crc\" DevicePath \"\"" Nov 24 09:06:37 crc kubenswrapper[4944]: I1124 09:06:37.501776 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" event={"ID":"a4e4656e-3009-4e6f-a478-c8cbefc40fba","Type":"ContainerDied","Data":"f6c39c0e22d53944b7fe4d7dc31ad226e0deb1c01152852fd9324cf576914ca3"} Nov 24 09:06:37 crc kubenswrapper[4944]: I1124 09:06:37.501811 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6c39c0e22d53944b7fe4d7dc31ad226e0deb1c01152852fd9324cf576914ca3" Nov 24 09:06:37 crc kubenswrapper[4944]: I1124 09:06:37.501842 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.766955 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g"] Nov 24 09:06:42 crc kubenswrapper[4944]: E1124 09:06:42.767991 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e4656e-3009-4e6f-a478-c8cbefc40fba" containerName="extract" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.768005 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e4656e-3009-4e6f-a478-c8cbefc40fba" containerName="extract" Nov 24 09:06:42 crc kubenswrapper[4944]: E1124 09:06:42.768015 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e4656e-3009-4e6f-a478-c8cbefc40fba" containerName="util" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.768021 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e4656e-3009-4e6f-a478-c8cbefc40fba" containerName="util" Nov 24 09:06:42 crc kubenswrapper[4944]: E1124 09:06:42.768039 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e4656e-3009-4e6f-a478-c8cbefc40fba" containerName="pull" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.768064 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e4656e-3009-4e6f-a478-c8cbefc40fba" containerName="pull" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.768160 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4e4656e-3009-4e6f-a478-c8cbefc40fba" containerName="extract" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.768562 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.770658 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.770680 4944 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-dv876" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.770820 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.832726 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g"] Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.882712 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/cc89a509-7a13-4557-bee8-39e3814044b8-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-z6x5g\" (UID: \"cc89a509-7a13-4557-bee8-39e3814044b8\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.882768 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7vqd\" (UniqueName: \"kubernetes.io/projected/cc89a509-7a13-4557-bee8-39e3814044b8-kube-api-access-g7vqd\") pod \"cert-manager-operator-controller-manager-64cf6dff88-z6x5g\" (UID: \"cc89a509-7a13-4557-bee8-39e3814044b8\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.985558 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/cc89a509-7a13-4557-bee8-39e3814044b8-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-z6x5g\" (UID: \"cc89a509-7a13-4557-bee8-39e3814044b8\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.985632 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7vqd\" (UniqueName: \"kubernetes.io/projected/cc89a509-7a13-4557-bee8-39e3814044b8-kube-api-access-g7vqd\") pod \"cert-manager-operator-controller-manager-64cf6dff88-z6x5g\" (UID: \"cc89a509-7a13-4557-bee8-39e3814044b8\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g" Nov 24 09:06:42 crc kubenswrapper[4944]: I1124 09:06:42.986252 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/cc89a509-7a13-4557-bee8-39e3814044b8-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-z6x5g\" (UID: \"cc89a509-7a13-4557-bee8-39e3814044b8\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g" Nov 24 09:06:43 crc kubenswrapper[4944]: I1124 09:06:43.011265 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7vqd\" (UniqueName: \"kubernetes.io/projected/cc89a509-7a13-4557-bee8-39e3814044b8-kube-api-access-g7vqd\") pod \"cert-manager-operator-controller-manager-64cf6dff88-z6x5g\" (UID: \"cc89a509-7a13-4557-bee8-39e3814044b8\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g" Nov 24 09:06:43 crc kubenswrapper[4944]: I1124 09:06:43.090316 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g" Nov 24 09:06:43 crc kubenswrapper[4944]: I1124 09:06:43.525557 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g"] Nov 24 09:06:43 crc kubenswrapper[4944]: I1124 09:06:43.539759 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g" event={"ID":"cc89a509-7a13-4557-bee8-39e3814044b8","Type":"ContainerStarted","Data":"810a00f44ac5ec402add160ab9f24578d5b3c118dc78166e9ad0681739ec5b37"} Nov 24 09:06:50 crc kubenswrapper[4944]: I1124 09:06:50.585894 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g" event={"ID":"cc89a509-7a13-4557-bee8-39e3814044b8","Type":"ContainerStarted","Data":"6f566053790a22e0a2207957f0778381e1a28d6141796fef9fd1589766e1874c"} Nov 24 09:06:50 crc kubenswrapper[4944]: I1124 09:06:50.608516 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-z6x5g" podStartSLOduration=1.712407213 podStartE2EDuration="8.608495638s" podCreationTimestamp="2025-11-24 09:06:42 +0000 UTC" firstStartedPulling="2025-11-24 09:06:43.535762521 +0000 UTC m=+864.070202983" lastFinishedPulling="2025-11-24 09:06:50.431850946 +0000 UTC m=+870.966291408" observedRunningTime="2025-11-24 09:06:50.600411658 +0000 UTC m=+871.134852120" watchObservedRunningTime="2025-11-24 09:06:50.608495638 +0000 UTC m=+871.142936100" Nov 24 09:06:53 crc kubenswrapper[4944]: I1124 09:06:53.548718 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:06:53 crc kubenswrapper[4944]: I1124 09:06:53.549118 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.066359 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-7gnzh"] Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.067118 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.068664 4944 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-cf6kr" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.068763 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.069554 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.077704 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-7gnzh"] Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.146801 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-726h7\" (UniqueName: \"kubernetes.io/projected/9aebbf05-0710-4819-97eb-cfb14cade6b3-kube-api-access-726h7\") pod \"cert-manager-webhook-f4fb5df64-7gnzh\" (UID: \"9aebbf05-0710-4819-97eb-cfb14cade6b3\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.146960 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9aebbf05-0710-4819-97eb-cfb14cade6b3-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-7gnzh\" (UID: \"9aebbf05-0710-4819-97eb-cfb14cade6b3\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.248785 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-726h7\" (UniqueName: \"kubernetes.io/projected/9aebbf05-0710-4819-97eb-cfb14cade6b3-kube-api-access-726h7\") pod \"cert-manager-webhook-f4fb5df64-7gnzh\" (UID: \"9aebbf05-0710-4819-97eb-cfb14cade6b3\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.248883 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9aebbf05-0710-4819-97eb-cfb14cade6b3-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-7gnzh\" (UID: \"9aebbf05-0710-4819-97eb-cfb14cade6b3\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.275799 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-726h7\" (UniqueName: \"kubernetes.io/projected/9aebbf05-0710-4819-97eb-cfb14cade6b3-kube-api-access-726h7\") pod \"cert-manager-webhook-f4fb5df64-7gnzh\" (UID: \"9aebbf05-0710-4819-97eb-cfb14cade6b3\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.276246 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9aebbf05-0710-4819-97eb-cfb14cade6b3-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-7gnzh\" (UID: \"9aebbf05-0710-4819-97eb-cfb14cade6b3\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.382401 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.804245 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv"] Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.807322 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.810336 4944 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-66r7x" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.811406 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv"] Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.866269 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-7gnzh"] Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.961067 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fd28b1a5-a8ce-4fc4-af42-3b3fc169e190-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-jt2wv\" (UID: \"fd28b1a5-a8ce-4fc4-af42-3b3fc169e190\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv" Nov 24 09:06:54 crc kubenswrapper[4944]: I1124 09:06:54.961146 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf8df\" (UniqueName: \"kubernetes.io/projected/fd28b1a5-a8ce-4fc4-af42-3b3fc169e190-kube-api-access-bf8df\") pod \"cert-manager-cainjector-855d9ccff4-jt2wv\" (UID: \"fd28b1a5-a8ce-4fc4-af42-3b3fc169e190\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv" Nov 24 09:06:55 crc kubenswrapper[4944]: I1124 09:06:55.062558 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fd28b1a5-a8ce-4fc4-af42-3b3fc169e190-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-jt2wv\" (UID: \"fd28b1a5-a8ce-4fc4-af42-3b3fc169e190\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv" Nov 24 09:06:55 crc kubenswrapper[4944]: I1124 09:06:55.062639 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf8df\" (UniqueName: \"kubernetes.io/projected/fd28b1a5-a8ce-4fc4-af42-3b3fc169e190-kube-api-access-bf8df\") pod \"cert-manager-cainjector-855d9ccff4-jt2wv\" (UID: \"fd28b1a5-a8ce-4fc4-af42-3b3fc169e190\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv" Nov 24 09:06:55 crc kubenswrapper[4944]: I1124 09:06:55.083734 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fd28b1a5-a8ce-4fc4-af42-3b3fc169e190-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-jt2wv\" (UID: \"fd28b1a5-a8ce-4fc4-af42-3b3fc169e190\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv" Nov 24 09:06:55 crc kubenswrapper[4944]: I1124 09:06:55.087280 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf8df\" (UniqueName: \"kubernetes.io/projected/fd28b1a5-a8ce-4fc4-af42-3b3fc169e190-kube-api-access-bf8df\") pod \"cert-manager-cainjector-855d9ccff4-jt2wv\" (UID: \"fd28b1a5-a8ce-4fc4-af42-3b3fc169e190\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv" Nov 24 09:06:55 crc kubenswrapper[4944]: I1124 09:06:55.124844 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv" Nov 24 09:06:55 crc kubenswrapper[4944]: I1124 09:06:55.566855 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv"] Nov 24 09:06:55 crc kubenswrapper[4944]: I1124 09:06:55.619586 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" event={"ID":"9aebbf05-0710-4819-97eb-cfb14cade6b3","Type":"ContainerStarted","Data":"a635e5d1f30514f2590ebe74096f49037e2ee5d260ff5b4a20b906149b793ba9"} Nov 24 09:06:55 crc kubenswrapper[4944]: I1124 09:06:55.620593 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv" event={"ID":"fd28b1a5-a8ce-4fc4-af42-3b3fc169e190","Type":"ContainerStarted","Data":"845dc57f6d5492d510c39ee31b57d949051e354a6640e39af25d12fcaf6a33ce"} Nov 24 09:07:04 crc kubenswrapper[4944]: I1124 09:07:04.703863 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" event={"ID":"9aebbf05-0710-4819-97eb-cfb14cade6b3","Type":"ContainerStarted","Data":"95b0e98a7b97787414108c5bee6b4e063c3150d385e87b086fdfcc37ab0a76c2"} Nov 24 09:07:04 crc kubenswrapper[4944]: I1124 09:07:04.704555 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" Nov 24 09:07:04 crc kubenswrapper[4944]: I1124 09:07:04.706886 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv" event={"ID":"fd28b1a5-a8ce-4fc4-af42-3b3fc169e190","Type":"ContainerStarted","Data":"9e3cc756709e7f4c5d8b5a1eaf8a037c3499536947c3a49d596969583dc22f16"} Nov 24 09:07:04 crc kubenswrapper[4944]: I1124 09:07:04.720029 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" podStartSLOduration=1.99780171 podStartE2EDuration="10.720014154s" podCreationTimestamp="2025-11-24 09:06:54 +0000 UTC" firstStartedPulling="2025-11-24 09:06:54.875997113 +0000 UTC m=+875.410437575" lastFinishedPulling="2025-11-24 09:07:03.598209557 +0000 UTC m=+884.132650019" observedRunningTime="2025-11-24 09:07:04.716706738 +0000 UTC m=+885.251147200" watchObservedRunningTime="2025-11-24 09:07:04.720014154 +0000 UTC m=+885.254454606" Nov 24 09:07:04 crc kubenswrapper[4944]: I1124 09:07:04.734167 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jt2wv" podStartSLOduration=2.678402579 podStartE2EDuration="10.734148078s" podCreationTimestamp="2025-11-24 09:06:54 +0000 UTC" firstStartedPulling="2025-11-24 09:06:55.577222654 +0000 UTC m=+876.111663116" lastFinishedPulling="2025-11-24 09:07:03.632968153 +0000 UTC m=+884.167408615" observedRunningTime="2025-11-24 09:07:04.729834419 +0000 UTC m=+885.264274891" watchObservedRunningTime="2025-11-24 09:07:04.734148078 +0000 UTC m=+885.268588540" Nov 24 09:07:09 crc kubenswrapper[4944]: I1124 09:07:09.386873 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-7gnzh" Nov 24 09:07:12 crc kubenswrapper[4944]: I1124 09:07:12.286772 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-p7czr"] Nov 24 09:07:12 crc kubenswrapper[4944]: I1124 09:07:12.288262 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-p7czr" Nov 24 09:07:12 crc kubenswrapper[4944]: I1124 09:07:12.291419 4944 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-5mnfw" Nov 24 09:07:12 crc kubenswrapper[4944]: I1124 09:07:12.298318 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-p7czr"] Nov 24 09:07:12 crc kubenswrapper[4944]: I1124 09:07:12.438002 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8bc6288c-a8c0-4615-9ff5-1c7839c75a77-bound-sa-token\") pod \"cert-manager-86cb77c54b-p7czr\" (UID: \"8bc6288c-a8c0-4615-9ff5-1c7839c75a77\") " pod="cert-manager/cert-manager-86cb77c54b-p7czr" Nov 24 09:07:12 crc kubenswrapper[4944]: I1124 09:07:12.438142 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh24f\" (UniqueName: \"kubernetes.io/projected/8bc6288c-a8c0-4615-9ff5-1c7839c75a77-kube-api-access-xh24f\") pod \"cert-manager-86cb77c54b-p7czr\" (UID: \"8bc6288c-a8c0-4615-9ff5-1c7839c75a77\") " pod="cert-manager/cert-manager-86cb77c54b-p7czr" Nov 24 09:07:12 crc kubenswrapper[4944]: I1124 09:07:12.540325 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh24f\" (UniqueName: \"kubernetes.io/projected/8bc6288c-a8c0-4615-9ff5-1c7839c75a77-kube-api-access-xh24f\") pod \"cert-manager-86cb77c54b-p7czr\" (UID: \"8bc6288c-a8c0-4615-9ff5-1c7839c75a77\") " pod="cert-manager/cert-manager-86cb77c54b-p7czr" Nov 24 09:07:12 crc kubenswrapper[4944]: I1124 09:07:12.540436 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8bc6288c-a8c0-4615-9ff5-1c7839c75a77-bound-sa-token\") pod \"cert-manager-86cb77c54b-p7czr\" (UID: \"8bc6288c-a8c0-4615-9ff5-1c7839c75a77\") " pod="cert-manager/cert-manager-86cb77c54b-p7czr" Nov 24 09:07:12 crc kubenswrapper[4944]: I1124 09:07:12.561200 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8bc6288c-a8c0-4615-9ff5-1c7839c75a77-bound-sa-token\") pod \"cert-manager-86cb77c54b-p7czr\" (UID: \"8bc6288c-a8c0-4615-9ff5-1c7839c75a77\") " pod="cert-manager/cert-manager-86cb77c54b-p7czr" Nov 24 09:07:12 crc kubenswrapper[4944]: I1124 09:07:12.561307 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh24f\" (UniqueName: \"kubernetes.io/projected/8bc6288c-a8c0-4615-9ff5-1c7839c75a77-kube-api-access-xh24f\") pod \"cert-manager-86cb77c54b-p7czr\" (UID: \"8bc6288c-a8c0-4615-9ff5-1c7839c75a77\") " pod="cert-manager/cert-manager-86cb77c54b-p7czr" Nov 24 09:07:12 crc kubenswrapper[4944]: I1124 09:07:12.639551 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-p7czr" Nov 24 09:07:13 crc kubenswrapper[4944]: I1124 09:07:13.082498 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-p7czr"] Nov 24 09:07:13 crc kubenswrapper[4944]: I1124 09:07:13.762734 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-p7czr" event={"ID":"8bc6288c-a8c0-4615-9ff5-1c7839c75a77","Type":"ContainerStarted","Data":"9223a611ae0d797a99e796b3544379b310cc7de17a854d04c6b0d684e55567d4"} Nov 24 09:07:13 crc kubenswrapper[4944]: I1124 09:07:13.763142 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-p7czr" event={"ID":"8bc6288c-a8c0-4615-9ff5-1c7839c75a77","Type":"ContainerStarted","Data":"873c2aa628719403324da17dc3a3d965cd5c69ea79c67e022e4de5680d021a71"} Nov 24 09:07:13 crc kubenswrapper[4944]: I1124 09:07:13.780486 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-p7czr" podStartSLOduration=1.78046513 podStartE2EDuration="1.78046513s" podCreationTimestamp="2025-11-24 09:07:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:07:13.775750349 +0000 UTC m=+894.310190821" watchObservedRunningTime="2025-11-24 09:07:13.78046513 +0000 UTC m=+894.314905592" Nov 24 09:07:23 crc kubenswrapper[4944]: I1124 09:07:23.550274 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:07:23 crc kubenswrapper[4944]: I1124 09:07:23.551035 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:07:24 crc kubenswrapper[4944]: I1124 09:07:24.995021 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zgw8d"] Nov 24 09:07:24 crc kubenswrapper[4944]: I1124 09:07:24.997805 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zgw8d" Nov 24 09:07:25 crc kubenswrapper[4944]: I1124 09:07:25.001689 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 09:07:25 crc kubenswrapper[4944]: I1124 09:07:25.001806 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zgw8d"] Nov 24 09:07:25 crc kubenswrapper[4944]: I1124 09:07:25.002252 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-8sfq2" Nov 24 09:07:25 crc kubenswrapper[4944]: I1124 09:07:25.005184 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 09:07:25 crc kubenswrapper[4944]: I1124 09:07:25.007298 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkfzh\" (UniqueName: \"kubernetes.io/projected/f2f1a239-2357-40b5-bf93-00ee35de0069-kube-api-access-fkfzh\") pod \"openstack-operator-index-zgw8d\" (UID: \"f2f1a239-2357-40b5-bf93-00ee35de0069\") " pod="openstack-operators/openstack-operator-index-zgw8d" Nov 24 09:07:25 crc kubenswrapper[4944]: I1124 09:07:25.108193 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkfzh\" (UniqueName: \"kubernetes.io/projected/f2f1a239-2357-40b5-bf93-00ee35de0069-kube-api-access-fkfzh\") pod \"openstack-operator-index-zgw8d\" (UID: \"f2f1a239-2357-40b5-bf93-00ee35de0069\") " pod="openstack-operators/openstack-operator-index-zgw8d" Nov 24 09:07:25 crc kubenswrapper[4944]: I1124 09:07:25.127722 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkfzh\" (UniqueName: \"kubernetes.io/projected/f2f1a239-2357-40b5-bf93-00ee35de0069-kube-api-access-fkfzh\") pod \"openstack-operator-index-zgw8d\" (UID: \"f2f1a239-2357-40b5-bf93-00ee35de0069\") " pod="openstack-operators/openstack-operator-index-zgw8d" Nov 24 09:07:25 crc kubenswrapper[4944]: I1124 09:07:25.315179 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zgw8d" Nov 24 09:07:25 crc kubenswrapper[4944]: I1124 09:07:25.700433 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zgw8d"] Nov 24 09:07:25 crc kubenswrapper[4944]: I1124 09:07:25.831105 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zgw8d" event={"ID":"f2f1a239-2357-40b5-bf93-00ee35de0069","Type":"ContainerStarted","Data":"f9eeb94658d5ce5e4fcc7d2c5b085bec1e0a2a09e999e277631dd2989ca3618f"} Nov 24 09:07:29 crc kubenswrapper[4944]: I1124 09:07:29.220687 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zgw8d" event={"ID":"f2f1a239-2357-40b5-bf93-00ee35de0069","Type":"ContainerStarted","Data":"186258cd690caa3197987fd89755001666e4c55c52b75eace9641b497974d79e"} Nov 24 09:07:29 crc kubenswrapper[4944]: I1124 09:07:29.240544 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zgw8d" podStartSLOduration=3.212879973 podStartE2EDuration="5.240520463s" podCreationTimestamp="2025-11-24 09:07:24 +0000 UTC" firstStartedPulling="2025-11-24 09:07:25.708157409 +0000 UTC m=+906.242597871" lastFinishedPulling="2025-11-24 09:07:27.735797899 +0000 UTC m=+908.270238361" observedRunningTime="2025-11-24 09:07:29.236863135 +0000 UTC m=+909.771303617" watchObservedRunningTime="2025-11-24 09:07:29.240520463 +0000 UTC m=+909.774960935" Nov 24 09:07:35 crc kubenswrapper[4944]: I1124 09:07:35.315464 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-zgw8d" Nov 24 09:07:35 crc kubenswrapper[4944]: I1124 09:07:35.315961 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-zgw8d" Nov 24 09:07:35 crc kubenswrapper[4944]: I1124 09:07:35.346707 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-zgw8d" Nov 24 09:07:36 crc kubenswrapper[4944]: I1124 09:07:36.291709 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-zgw8d" Nov 24 09:07:38 crc kubenswrapper[4944]: I1124 09:07:38.826798 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm"] Nov 24 09:07:38 crc kubenswrapper[4944]: I1124 09:07:38.828574 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:38 crc kubenswrapper[4944]: I1124 09:07:38.831173 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-fphmk" Nov 24 09:07:38 crc kubenswrapper[4944]: I1124 09:07:38.839617 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm"] Nov 24 09:07:39 crc kubenswrapper[4944]: I1124 09:07:39.030714 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scfhj\" (UniqueName: \"kubernetes.io/projected/4ed11819-58d0-45ab-8fd6-3930861c2bc9-kube-api-access-scfhj\") pod \"eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:39 crc kubenswrapper[4944]: I1124 09:07:39.030778 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-util\") pod \"eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:39 crc kubenswrapper[4944]: I1124 09:07:39.030850 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-bundle\") pod \"eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:39 crc kubenswrapper[4944]: I1124 09:07:39.132216 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scfhj\" (UniqueName: \"kubernetes.io/projected/4ed11819-58d0-45ab-8fd6-3930861c2bc9-kube-api-access-scfhj\") pod \"eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:39 crc kubenswrapper[4944]: I1124 09:07:39.132521 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-util\") pod \"eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:39 crc kubenswrapper[4944]: I1124 09:07:39.132608 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-bundle\") pod \"eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:39 crc kubenswrapper[4944]: I1124 09:07:39.133308 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-bundle\") pod \"eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:39 crc kubenswrapper[4944]: I1124 09:07:39.133487 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-util\") pod \"eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:39 crc kubenswrapper[4944]: I1124 09:07:39.155039 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scfhj\" (UniqueName: \"kubernetes.io/projected/4ed11819-58d0-45ab-8fd6-3930861c2bc9-kube-api-access-scfhj\") pod \"eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:39 crc kubenswrapper[4944]: I1124 09:07:39.448732 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:39 crc kubenswrapper[4944]: I1124 09:07:39.657719 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm"] Nov 24 09:07:39 crc kubenswrapper[4944]: W1124 09:07:39.666093 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ed11819_58d0_45ab_8fd6_3930861c2bc9.slice/crio-5708a53b0a47e2ae7be8430be54f4feb344645d04c68790fd545885b6b91af85 WatchSource:0}: Error finding container 5708a53b0a47e2ae7be8430be54f4feb344645d04c68790fd545885b6b91af85: Status 404 returned error can't find the container with id 5708a53b0a47e2ae7be8430be54f4feb344645d04c68790fd545885b6b91af85 Nov 24 09:07:40 crc kubenswrapper[4944]: I1124 09:07:40.286691 4944 generic.go:334] "Generic (PLEG): container finished" podID="4ed11819-58d0-45ab-8fd6-3930861c2bc9" containerID="8f782c9a58c87f58a271b364abd564befd99164a0ea71c5a77966272913b7ad0" exitCode=0 Nov 24 09:07:40 crc kubenswrapper[4944]: I1124 09:07:40.286802 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" event={"ID":"4ed11819-58d0-45ab-8fd6-3930861c2bc9","Type":"ContainerDied","Data":"8f782c9a58c87f58a271b364abd564befd99164a0ea71c5a77966272913b7ad0"} Nov 24 09:07:40 crc kubenswrapper[4944]: I1124 09:07:40.287170 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" event={"ID":"4ed11819-58d0-45ab-8fd6-3930861c2bc9","Type":"ContainerStarted","Data":"5708a53b0a47e2ae7be8430be54f4feb344645d04c68790fd545885b6b91af85"} Nov 24 09:07:41 crc kubenswrapper[4944]: I1124 09:07:41.296204 4944 generic.go:334] "Generic (PLEG): container finished" podID="4ed11819-58d0-45ab-8fd6-3930861c2bc9" containerID="f7c5394c4ec1b3ab9855dab84065abb5e1d28a1226878e3ab292ebde342d84f2" exitCode=0 Nov 24 09:07:41 crc kubenswrapper[4944]: I1124 09:07:41.296257 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" event={"ID":"4ed11819-58d0-45ab-8fd6-3930861c2bc9","Type":"ContainerDied","Data":"f7c5394c4ec1b3ab9855dab84065abb5e1d28a1226878e3ab292ebde342d84f2"} Nov 24 09:07:42 crc kubenswrapper[4944]: I1124 09:07:42.302485 4944 generic.go:334] "Generic (PLEG): container finished" podID="4ed11819-58d0-45ab-8fd6-3930861c2bc9" containerID="ae4d4c3a68c573271305acd6b91d706ac0c67a9e037980e999a031da63e97e53" exitCode=0 Nov 24 09:07:42 crc kubenswrapper[4944]: I1124 09:07:42.302685 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" event={"ID":"4ed11819-58d0-45ab-8fd6-3930861c2bc9","Type":"ContainerDied","Data":"ae4d4c3a68c573271305acd6b91d706ac0c67a9e037980e999a031da63e97e53"} Nov 24 09:07:43 crc kubenswrapper[4944]: I1124 09:07:43.546821 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:43 crc kubenswrapper[4944]: I1124 09:07:43.690880 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-bundle\") pod \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " Nov 24 09:07:43 crc kubenswrapper[4944]: I1124 09:07:43.690988 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-util\") pod \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " Nov 24 09:07:43 crc kubenswrapper[4944]: I1124 09:07:43.691026 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scfhj\" (UniqueName: \"kubernetes.io/projected/4ed11819-58d0-45ab-8fd6-3930861c2bc9-kube-api-access-scfhj\") pod \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\" (UID: \"4ed11819-58d0-45ab-8fd6-3930861c2bc9\") " Nov 24 09:07:43 crc kubenswrapper[4944]: I1124 09:07:43.692040 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-bundle" (OuterVolumeSpecName: "bundle") pod "4ed11819-58d0-45ab-8fd6-3930861c2bc9" (UID: "4ed11819-58d0-45ab-8fd6-3930861c2bc9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:07:43 crc kubenswrapper[4944]: I1124 09:07:43.695882 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ed11819-58d0-45ab-8fd6-3930861c2bc9-kube-api-access-scfhj" (OuterVolumeSpecName: "kube-api-access-scfhj") pod "4ed11819-58d0-45ab-8fd6-3930861c2bc9" (UID: "4ed11819-58d0-45ab-8fd6-3930861c2bc9"). InnerVolumeSpecName "kube-api-access-scfhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:07:43 crc kubenswrapper[4944]: I1124 09:07:43.704940 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-util" (OuterVolumeSpecName: "util") pod "4ed11819-58d0-45ab-8fd6-3930861c2bc9" (UID: "4ed11819-58d0-45ab-8fd6-3930861c2bc9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:07:43 crc kubenswrapper[4944]: I1124 09:07:43.792351 4944 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:43 crc kubenswrapper[4944]: I1124 09:07:43.792393 4944 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ed11819-58d0-45ab-8fd6-3930861c2bc9-util\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:43 crc kubenswrapper[4944]: I1124 09:07:43.792404 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scfhj\" (UniqueName: \"kubernetes.io/projected/4ed11819-58d0-45ab-8fd6-3930861c2bc9-kube-api-access-scfhj\") on node \"crc\" DevicePath \"\"" Nov 24 09:07:44 crc kubenswrapper[4944]: I1124 09:07:44.317462 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" event={"ID":"4ed11819-58d0-45ab-8fd6-3930861c2bc9","Type":"ContainerDied","Data":"5708a53b0a47e2ae7be8430be54f4feb344645d04c68790fd545885b6b91af85"} Nov 24 09:07:44 crc kubenswrapper[4944]: I1124 09:07:44.317512 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5708a53b0a47e2ae7be8430be54f4feb344645d04c68790fd545885b6b91af85" Nov 24 09:07:44 crc kubenswrapper[4944]: I1124 09:07:44.317527 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm" Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.268318 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4"] Nov 24 09:07:50 crc kubenswrapper[4944]: E1124 09:07:50.270155 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed11819-58d0-45ab-8fd6-3930861c2bc9" containerName="util" Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.270237 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed11819-58d0-45ab-8fd6-3930861c2bc9" containerName="util" Nov 24 09:07:50 crc kubenswrapper[4944]: E1124 09:07:50.270313 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed11819-58d0-45ab-8fd6-3930861c2bc9" containerName="extract" Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.270369 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed11819-58d0-45ab-8fd6-3930861c2bc9" containerName="extract" Nov 24 09:07:50 crc kubenswrapper[4944]: E1124 09:07:50.270421 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed11819-58d0-45ab-8fd6-3930861c2bc9" containerName="pull" Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.270475 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed11819-58d0-45ab-8fd6-3930861c2bc9" containerName="pull" Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.270620 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ed11819-58d0-45ab-8fd6-3930861c2bc9" containerName="extract" Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.271318 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4" Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.273744 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-fh9wm" Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.300495 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4"] Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.374356 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnzl2\" (UniqueName: \"kubernetes.io/projected/35d7929f-4374-4c32-826a-f0dc8df3d8ff-kube-api-access-wnzl2\") pod \"openstack-operator-controller-operator-56cb4fc9f6-t2lj4\" (UID: \"35d7929f-4374-4c32-826a-f0dc8df3d8ff\") " pod="openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4" Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.475484 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnzl2\" (UniqueName: \"kubernetes.io/projected/35d7929f-4374-4c32-826a-f0dc8df3d8ff-kube-api-access-wnzl2\") pod \"openstack-operator-controller-operator-56cb4fc9f6-t2lj4\" (UID: \"35d7929f-4374-4c32-826a-f0dc8df3d8ff\") " pod="openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4" Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.497606 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnzl2\" (UniqueName: \"kubernetes.io/projected/35d7929f-4374-4c32-826a-f0dc8df3d8ff-kube-api-access-wnzl2\") pod \"openstack-operator-controller-operator-56cb4fc9f6-t2lj4\" (UID: \"35d7929f-4374-4c32-826a-f0dc8df3d8ff\") " pod="openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4" Nov 24 09:07:50 crc kubenswrapper[4944]: I1124 09:07:50.592524 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4" Nov 24 09:07:51 crc kubenswrapper[4944]: I1124 09:07:51.050865 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4"] Nov 24 09:07:51 crc kubenswrapper[4944]: I1124 09:07:51.408708 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4" event={"ID":"35d7929f-4374-4c32-826a-f0dc8df3d8ff","Type":"ContainerStarted","Data":"fe5c193e777d3be72cd0d94229af552b374360f391021650800d2e25c1b4f94d"} Nov 24 09:07:53 crc kubenswrapper[4944]: I1124 09:07:53.548379 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:07:53 crc kubenswrapper[4944]: I1124 09:07:53.548799 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:07:53 crc kubenswrapper[4944]: I1124 09:07:53.548846 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:07:53 crc kubenswrapper[4944]: I1124 09:07:53.549496 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b51cb7736616a4e1de4d8a3f423bb2e041c1d8aa22c653e62b3b1a70f4ce6afd"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:07:53 crc kubenswrapper[4944]: I1124 09:07:53.549571 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://b51cb7736616a4e1de4d8a3f423bb2e041c1d8aa22c653e62b3b1a70f4ce6afd" gracePeriod=600 Nov 24 09:07:54 crc kubenswrapper[4944]: I1124 09:07:54.430899 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="b51cb7736616a4e1de4d8a3f423bb2e041c1d8aa22c653e62b3b1a70f4ce6afd" exitCode=0 Nov 24 09:07:54 crc kubenswrapper[4944]: I1124 09:07:54.430967 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"b51cb7736616a4e1de4d8a3f423bb2e041c1d8aa22c653e62b3b1a70f4ce6afd"} Nov 24 09:07:54 crc kubenswrapper[4944]: I1124 09:07:54.431068 4944 scope.go:117] "RemoveContainer" containerID="f61d46fb34a7fb69e551ab18c2960366dce6b114b94216eab6db96b03be79c6f" Nov 24 09:07:55 crc kubenswrapper[4944]: I1124 09:07:55.438702 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"1d4676a5ed72cda8e3ecd73f73356ca831b52b82f176ed2715695ec66c7c950e"} Nov 24 09:07:55 crc kubenswrapper[4944]: I1124 09:07:55.440725 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4" event={"ID":"35d7929f-4374-4c32-826a-f0dc8df3d8ff","Type":"ContainerStarted","Data":"b6ea2a08a1248631f0c0fdc0ce2f87553452fb20482b9f3bf4f77a54cadf2711"} Nov 24 09:07:58 crc kubenswrapper[4944]: I1124 09:07:58.464872 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4" event={"ID":"35d7929f-4374-4c32-826a-f0dc8df3d8ff","Type":"ContainerStarted","Data":"ce61d63b70916aa083014fb7983e285d13f926015bd4c49b8d5ad8c3b0c8438b"} Nov 24 09:07:58 crc kubenswrapper[4944]: I1124 09:07:58.466960 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4" Nov 24 09:07:58 crc kubenswrapper[4944]: I1124 09:07:58.535273 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4" podStartSLOduration=2.181630473 podStartE2EDuration="8.535254005s" podCreationTimestamp="2025-11-24 09:07:50 +0000 UTC" firstStartedPulling="2025-11-24 09:07:51.070019082 +0000 UTC m=+931.604459544" lastFinishedPulling="2025-11-24 09:07:57.423642614 +0000 UTC m=+937.958083076" observedRunningTime="2025-11-24 09:07:58.530170581 +0000 UTC m=+939.064611073" watchObservedRunningTime="2025-11-24 09:07:58.535254005 +0000 UTC m=+939.069694467" Nov 24 09:08:00 crc kubenswrapper[4944]: I1124 09:08:00.477586 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-56cb4fc9f6-t2lj4" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.850302 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7"] Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.852273 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.856084 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g"] Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.857214 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-w92fl" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.857587 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.858966 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-56876" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.865078 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7"] Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.886137 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g"] Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.894332 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg"] Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.895589 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.897863 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-d4rr9" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.909539 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldk5b\" (UniqueName: \"kubernetes.io/projected/28277a17-6dab-4f8e-b765-46fefba90b01-kube-api-access-ldk5b\") pod \"designate-operator-controller-manager-767ccfd65f-864hg\" (UID: \"28277a17-6dab-4f8e-b765-46fefba90b01\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.909699 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb8c7\" (UniqueName: \"kubernetes.io/projected/2640bc37-2c54-4e38-ae79-a1d5cc08af65-kube-api-access-bb8c7\") pod \"cinder-operator-controller-manager-6498cbf48f-cpf2g\" (UID: \"2640bc37-2c54-4e38-ae79-a1d5cc08af65\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.909727 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2mmr\" (UniqueName: \"kubernetes.io/projected/f2b53031-1259-46ff-a42d-f65547877116-kube-api-access-f2mmr\") pod \"barbican-operator-controller-manager-75fb479bcc-qvqs7\" (UID: \"f2b53031-1259-46ff-a42d-f65547877116\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.912226 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg"] Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.916260 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn"] Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.921355 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.923650 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-nrzh9" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.923845 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn"] Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.942640 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k"] Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.943951 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.952177 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-rgbjw" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.971259 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k"] Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.983326 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk"] Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.984316 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk" Nov 24 09:08:18 crc kubenswrapper[4944]: I1124 09:08:18.996463 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-nnmxv" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.007192 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.008806 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.014117 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb8c7\" (UniqueName: \"kubernetes.io/projected/2640bc37-2c54-4e38-ae79-a1d5cc08af65-kube-api-access-bb8c7\") pod \"cinder-operator-controller-manager-6498cbf48f-cpf2g\" (UID: \"2640bc37-2c54-4e38-ae79-a1d5cc08af65\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.014163 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqfdk\" (UniqueName: \"kubernetes.io/projected/916de258-fc43-4fcc-bd66-9e24e3216ac8-kube-api-access-bqfdk\") pod \"glance-operator-controller-manager-7969689c84-8b2jn\" (UID: \"916de258-fc43-4fcc-bd66-9e24e3216ac8\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.014216 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2mmr\" (UniqueName: \"kubernetes.io/projected/f2b53031-1259-46ff-a42d-f65547877116-kube-api-access-f2mmr\") pod \"barbican-operator-controller-manager-75fb479bcc-qvqs7\" (UID: \"f2b53031-1259-46ff-a42d-f65547877116\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.014295 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7jvp\" (UniqueName: \"kubernetes.io/projected/655196aa-961a-4b69-8fad-6975c0ce44f2-kube-api-access-w7jvp\") pod \"horizon-operator-controller-manager-598f69df5d-bnxrk\" (UID: \"655196aa-961a-4b69-8fad-6975c0ce44f2\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.014357 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldk5b\" (UniqueName: \"kubernetes.io/projected/28277a17-6dab-4f8e-b765-46fefba90b01-kube-api-access-ldk5b\") pod \"designate-operator-controller-manager-767ccfd65f-864hg\" (UID: \"28277a17-6dab-4f8e-b765-46fefba90b01\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.014477 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59sx6\" (UniqueName: \"kubernetes.io/projected/265e80c3-fc9b-43f0-a8f8-6f3a4562a9ca-kube-api-access-59sx6\") pod \"heat-operator-controller-manager-56f54d6746-pq27k\" (UID: \"265e80c3-fc9b-43f0-a8f8-6f3a4562a9ca\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.015059 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.015248 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-crgbg" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.033772 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.044832 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.046232 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.050661 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-5n9fb" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.062407 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2mmr\" (UniqueName: \"kubernetes.io/projected/f2b53031-1259-46ff-a42d-f65547877116-kube-api-access-f2mmr\") pod \"barbican-operator-controller-manager-75fb479bcc-qvqs7\" (UID: \"f2b53031-1259-46ff-a42d-f65547877116\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.065718 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.067691 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldk5b\" (UniqueName: \"kubernetes.io/projected/28277a17-6dab-4f8e-b765-46fefba90b01-kube-api-access-ldk5b\") pod \"designate-operator-controller-manager-767ccfd65f-864hg\" (UID: \"28277a17-6dab-4f8e-b765-46fefba90b01\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.072757 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb8c7\" (UniqueName: \"kubernetes.io/projected/2640bc37-2c54-4e38-ae79-a1d5cc08af65-kube-api-access-bb8c7\") pod \"cinder-operator-controller-manager-6498cbf48f-cpf2g\" (UID: \"2640bc37-2c54-4e38-ae79-a1d5cc08af65\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.090335 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.108650 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.109848 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.115250 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-82vxv" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.116540 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqfdk\" (UniqueName: \"kubernetes.io/projected/916de258-fc43-4fcc-bd66-9e24e3216ac8-kube-api-access-bqfdk\") pod \"glance-operator-controller-manager-7969689c84-8b2jn\" (UID: \"916de258-fc43-4fcc-bd66-9e24e3216ac8\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.116577 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qnjv\" (UniqueName: \"kubernetes.io/projected/f5f1851e-a78f-458f-9c7a-fa271cfa4283-kube-api-access-7qnjv\") pod \"ironic-operator-controller-manager-99b499f4-jq6w6\" (UID: \"f5f1851e-a78f-458f-9c7a-fa271cfa4283\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.116626 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26e6c9ac-040e-4e10-97ca-6000969950f3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-gvzfc\" (UID: \"26e6c9ac-040e-4e10-97ca-6000969950f3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.116652 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7jvp\" (UniqueName: \"kubernetes.io/projected/655196aa-961a-4b69-8fad-6975c0ce44f2-kube-api-access-w7jvp\") pod \"horizon-operator-controller-manager-598f69df5d-bnxrk\" (UID: \"655196aa-961a-4b69-8fad-6975c0ce44f2\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.116680 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hmlm\" (UniqueName: \"kubernetes.io/projected/26e6c9ac-040e-4e10-97ca-6000969950f3-kube-api-access-5hmlm\") pod \"infra-operator-controller-manager-6dd8864d7c-gvzfc\" (UID: \"26e6c9ac-040e-4e10-97ca-6000969950f3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.116709 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59sx6\" (UniqueName: \"kubernetes.io/projected/265e80c3-fc9b-43f0-a8f8-6f3a4562a9ca-kube-api-access-59sx6\") pod \"heat-operator-controller-manager-56f54d6746-pq27k\" (UID: \"265e80c3-fc9b-43f0-a8f8-6f3a4562a9ca\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.117211 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.118182 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.128301 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-nth9w" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.136347 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.137349 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.139829 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-4lsrc" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.148178 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7jvp\" (UniqueName: \"kubernetes.io/projected/655196aa-961a-4b69-8fad-6975c0ce44f2-kube-api-access-w7jvp\") pod \"horizon-operator-controller-manager-598f69df5d-bnxrk\" (UID: \"655196aa-961a-4b69-8fad-6975c0ce44f2\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.148244 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.150483 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqfdk\" (UniqueName: \"kubernetes.io/projected/916de258-fc43-4fcc-bd66-9e24e3216ac8-kube-api-access-bqfdk\") pod \"glance-operator-controller-manager-7969689c84-8b2jn\" (UID: \"916de258-fc43-4fcc-bd66-9e24e3216ac8\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.155124 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.158778 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59sx6\" (UniqueName: \"kubernetes.io/projected/265e80c3-fc9b-43f0-a8f8-6f3a4562a9ca-kube-api-access-59sx6\") pod \"heat-operator-controller-manager-56f54d6746-pq27k\" (UID: \"265e80c3-fc9b-43f0-a8f8-6f3a4562a9ca\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.158859 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.160083 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.162716 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-l5qgd" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.162902 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.164069 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.166847 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-9fqtj" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.176784 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.180813 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.187107 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.191065 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.198213 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.222098 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.224195 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.258760 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.309538 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.312087 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qnjv\" (UniqueName: \"kubernetes.io/projected/f5f1851e-a78f-458f-9c7a-fa271cfa4283-kube-api-access-7qnjv\") pod \"ironic-operator-controller-manager-99b499f4-jq6w6\" (UID: \"f5f1851e-a78f-458f-9c7a-fa271cfa4283\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.312240 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5gcx\" (UniqueName: \"kubernetes.io/projected/163609af-d6a1-48b2-a5cb-e844edff0ef4-kube-api-access-j5gcx\") pod \"nova-operator-controller-manager-cfbb9c588-p4tcl\" (UID: \"163609af-d6a1-48b2-a5cb-e844edff0ef4\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.312416 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26e6c9ac-040e-4e10-97ca-6000969950f3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-gvzfc\" (UID: \"26e6c9ac-040e-4e10-97ca-6000969950f3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.312525 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkknr\" (UniqueName: \"kubernetes.io/projected/3e0e56a5-c24a-4dea-b478-d742409c80b4-kube-api-access-bkknr\") pod \"neutron-operator-controller-manager-78bd47f458-td8rp\" (UID: \"3e0e56a5-c24a-4dea-b478-d742409c80b4\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.312637 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdpl9\" (UniqueName: \"kubernetes.io/projected/f62b2a83-584b-4f8a-aa62-c1e7499100bd-kube-api-access-kdpl9\") pod \"manila-operator-controller-manager-58f887965d-d5mp6\" (UID: \"f62b2a83-584b-4f8a-aa62-c1e7499100bd\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.312719 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hmlm\" (UniqueName: \"kubernetes.io/projected/26e6c9ac-040e-4e10-97ca-6000969950f3-kube-api-access-5hmlm\") pod \"infra-operator-controller-manager-6dd8864d7c-gvzfc\" (UID: \"26e6c9ac-040e-4e10-97ca-6000969950f3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.312806 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdcxj\" (UniqueName: \"kubernetes.io/projected/ec7ed47f-2c2c-4401-ac0a-0510d8649f01-kube-api-access-bdcxj\") pod \"keystone-operator-controller-manager-7454b96578-wgm4r\" (UID: \"ec7ed47f-2c2c-4401-ac0a-0510d8649f01\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.312934 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72ff8\" (UniqueName: \"kubernetes.io/projected/fe8abc9f-e87f-4417-905f-44bb0b432a59-kube-api-access-72ff8\") pod \"mariadb-operator-controller-manager-54b5986bb8-wjthz\" (UID: \"fe8abc9f-e87f-4417-905f-44bb0b432a59\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.313796 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.314589 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk" Nov 24 09:08:19 crc kubenswrapper[4944]: E1124 09:08:19.315145 4944 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 09:08:19 crc kubenswrapper[4944]: E1124 09:08:19.315277 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/26e6c9ac-040e-4e10-97ca-6000969950f3-cert podName:26e6c9ac-040e-4e10-97ca-6000969950f3 nodeName:}" failed. No retries permitted until 2025-11-24 09:08:19.815257155 +0000 UTC m=+960.349697617 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/26e6c9ac-040e-4e10-97ca-6000969950f3-cert") pod "infra-operator-controller-manager-6dd8864d7c-gvzfc" (UID: "26e6c9ac-040e-4e10-97ca-6000969950f3") : secret "infra-operator-webhook-server-cert" not found Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.316499 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-mj2tf" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.337814 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.348087 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.349570 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.350641 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qnjv\" (UniqueName: \"kubernetes.io/projected/f5f1851e-a78f-458f-9c7a-fa271cfa4283-kube-api-access-7qnjv\") pod \"ironic-operator-controller-manager-99b499f4-jq6w6\" (UID: \"f5f1851e-a78f-458f-9c7a-fa271cfa4283\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.355240 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-88z72" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.355255 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hmlm\" (UniqueName: \"kubernetes.io/projected/26e6c9ac-040e-4e10-97ca-6000969950f3-kube-api-access-5hmlm\") pod \"infra-operator-controller-manager-6dd8864d7c-gvzfc\" (UID: \"26e6c9ac-040e-4e10-97ca-6000969950f3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.369332 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.392530 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.394014 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.397183 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xswz5" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.399234 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.414465 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkknr\" (UniqueName: \"kubernetes.io/projected/3e0e56a5-c24a-4dea-b478-d742409c80b4-kube-api-access-bkknr\") pod \"neutron-operator-controller-manager-78bd47f458-td8rp\" (UID: \"3e0e56a5-c24a-4dea-b478-d742409c80b4\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.414522 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdpl9\" (UniqueName: \"kubernetes.io/projected/f62b2a83-584b-4f8a-aa62-c1e7499100bd-kube-api-access-kdpl9\") pod \"manila-operator-controller-manager-58f887965d-d5mp6\" (UID: \"f62b2a83-584b-4f8a-aa62-c1e7499100bd\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.414553 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdcxj\" (UniqueName: \"kubernetes.io/projected/ec7ed47f-2c2c-4401-ac0a-0510d8649f01-kube-api-access-bdcxj\") pod \"keystone-operator-controller-manager-7454b96578-wgm4r\" (UID: \"ec7ed47f-2c2c-4401-ac0a-0510d8649f01\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.414596 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6qtr\" (UniqueName: \"kubernetes.io/projected/c702883d-7a93-4ede-a1dc-b2becaf94999-kube-api-access-v6qtr\") pod \"octavia-operator-controller-manager-54cfbf4c7d-gcggn\" (UID: \"c702883d-7a93-4ede-a1dc-b2becaf94999\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.414629 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72ff8\" (UniqueName: \"kubernetes.io/projected/fe8abc9f-e87f-4417-905f-44bb0b432a59-kube-api-access-72ff8\") pod \"mariadb-operator-controller-manager-54b5986bb8-wjthz\" (UID: \"fe8abc9f-e87f-4417-905f-44bb0b432a59\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.414693 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5gcx\" (UniqueName: \"kubernetes.io/projected/163609af-d6a1-48b2-a5cb-e844edff0ef4-kube-api-access-j5gcx\") pod \"nova-operator-controller-manager-cfbb9c588-p4tcl\" (UID: \"163609af-d6a1-48b2-a5cb-e844edff0ef4\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.415210 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.415863 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.416900 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.419656 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.422463 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-fnm6h" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.426688 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.432310 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.433670 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.434719 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.436414 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.452893 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.457111 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.471128 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.472422 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.478143 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.479199 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-qmvgd" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.479375 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-49v5j" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.480263 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5gcx\" (UniqueName: \"kubernetes.io/projected/163609af-d6a1-48b2-a5cb-e844edff0ef4-kube-api-access-j5gcx\") pod \"nova-operator-controller-manager-cfbb9c588-p4tcl\" (UID: \"163609af-d6a1-48b2-a5cb-e844edff0ef4\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.480749 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdpl9\" (UniqueName: \"kubernetes.io/projected/f62b2a83-584b-4f8a-aa62-c1e7499100bd-kube-api-access-kdpl9\") pod \"manila-operator-controller-manager-58f887965d-d5mp6\" (UID: \"f62b2a83-584b-4f8a-aa62-c1e7499100bd\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.482092 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bs9j2" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.488841 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72ff8\" (UniqueName: \"kubernetes.io/projected/fe8abc9f-e87f-4417-905f-44bb0b432a59-kube-api-access-72ff8\") pod \"mariadb-operator-controller-manager-54b5986bb8-wjthz\" (UID: \"fe8abc9f-e87f-4417-905f-44bb0b432a59\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.505484 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdcxj\" (UniqueName: \"kubernetes.io/projected/ec7ed47f-2c2c-4401-ac0a-0510d8649f01-kube-api-access-bdcxj\") pod \"keystone-operator-controller-manager-7454b96578-wgm4r\" (UID: \"ec7ed47f-2c2c-4401-ac0a-0510d8649f01\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.505961 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.517114 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.518533 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.519677 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs68g\" (UniqueName: \"kubernetes.io/projected/6f6767fe-2e4b-4b7a-b92c-8007220e1f3e-kube-api-access-fs68g\") pod \"ovn-operator-controller-manager-54fc5f65b7-xblqz\" (UID: \"6f6767fe-2e4b-4b7a-b92c-8007220e1f3e\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.519728 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6qtr\" (UniqueName: \"kubernetes.io/projected/c702883d-7a93-4ede-a1dc-b2becaf94999-kube-api-access-v6qtr\") pod \"octavia-operator-controller-manager-54cfbf4c7d-gcggn\" (UID: \"c702883d-7a93-4ede-a1dc-b2becaf94999\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.519796 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8jcm\" (UniqueName: \"kubernetes.io/projected/64f7e2cc-d596-4270-863c-0fa67797d789-kube-api-access-f8jcm\") pod \"placement-operator-controller-manager-5b797b8dff-dvc87\" (UID: \"64f7e2cc-d596-4270-863c-0fa67797d789\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.519826 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42f09610-472c-44b1-85d4-55fa62d2c2c1-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf\" (UID: \"42f09610-472c-44b1-85d4-55fa62d2c2c1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.519861 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmm4m\" (UniqueName: \"kubernetes.io/projected/42f09610-472c-44b1-85d4-55fa62d2c2c1-kube-api-access-fmm4m\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf\" (UID: \"42f09610-472c-44b1-85d4-55fa62d2c2c1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.528908 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.558012 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkknr\" (UniqueName: \"kubernetes.io/projected/3e0e56a5-c24a-4dea-b478-d742409c80b4-kube-api-access-bkknr\") pod \"neutron-operator-controller-manager-78bd47f458-td8rp\" (UID: \"3e0e56a5-c24a-4dea-b478-d742409c80b4\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.559204 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-s2ttb" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.600759 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6qtr\" (UniqueName: \"kubernetes.io/projected/c702883d-7a93-4ede-a1dc-b2becaf94999-kube-api-access-v6qtr\") pod \"octavia-operator-controller-manager-54cfbf4c7d-gcggn\" (UID: \"c702883d-7a93-4ede-a1dc-b2becaf94999\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.601083 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.623664 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8jcm\" (UniqueName: \"kubernetes.io/projected/64f7e2cc-d596-4270-863c-0fa67797d789-kube-api-access-f8jcm\") pod \"placement-operator-controller-manager-5b797b8dff-dvc87\" (UID: \"64f7e2cc-d596-4270-863c-0fa67797d789\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.623730 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42f09610-472c-44b1-85d4-55fa62d2c2c1-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf\" (UID: \"42f09610-472c-44b1-85d4-55fa62d2c2c1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.623767 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gvjs\" (UniqueName: \"kubernetes.io/projected/fed38002-b5d5-4bc0-9fde-5cb9e04c5542-kube-api-access-2gvjs\") pod \"swift-operator-controller-manager-d656998f4-dlfx8\" (UID: \"fed38002-b5d5-4bc0-9fde-5cb9e04c5542\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.623805 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpzkd\" (UniqueName: \"kubernetes.io/projected/82fc1284-da3b-4620-ae81-64535b5dde90-kube-api-access-tpzkd\") pod \"watcher-operator-controller-manager-8c6448b9f-54vjr\" (UID: \"82fc1284-da3b-4620-ae81-64535b5dde90\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.623837 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpcgd\" (UniqueName: \"kubernetes.io/projected/63a16eba-7d99-41f4-9c77-83a8497c4210-kube-api-access-tpcgd\") pod \"test-operator-controller-manager-b4c496f69-j5gtk\" (UID: \"63a16eba-7d99-41f4-9c77-83a8497c4210\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.623870 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmm4m\" (UniqueName: \"kubernetes.io/projected/42f09610-472c-44b1-85d4-55fa62d2c2c1-kube-api-access-fmm4m\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf\" (UID: \"42f09610-472c-44b1-85d4-55fa62d2c2c1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.623954 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs68g\" (UniqueName: \"kubernetes.io/projected/6f6767fe-2e4b-4b7a-b92c-8007220e1f3e-kube-api-access-fs68g\") pod \"ovn-operator-controller-manager-54fc5f65b7-xblqz\" (UID: \"6f6767fe-2e4b-4b7a-b92c-8007220e1f3e\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.624009 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stx4d\" (UniqueName: \"kubernetes.io/projected/b4532ccb-d13d-4cd8-9ccb-240cb2d84195-kube-api-access-stx4d\") pod \"telemetry-operator-controller-manager-6d4bf84b58-2h8kk\" (UID: \"b4532ccb-d13d-4cd8-9ccb-240cb2d84195\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.627365 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp" Nov 24 09:08:19 crc kubenswrapper[4944]: E1124 09:08:19.628063 4944 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 09:08:19 crc kubenswrapper[4944]: E1124 09:08:19.628242 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/42f09610-472c-44b1-85d4-55fa62d2c2c1-cert podName:42f09610-472c-44b1-85d4-55fa62d2c2c1 nodeName:}" failed. No retries permitted until 2025-11-24 09:08:20.128219666 +0000 UTC m=+960.662660118 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/42f09610-472c-44b1-85d4-55fa62d2c2c1-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" (UID: "42f09610-472c-44b1-85d4-55fa62d2c2c1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.648883 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.656436 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.662366 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8jcm\" (UniqueName: \"kubernetes.io/projected/64f7e2cc-d596-4270-863c-0fa67797d789-kube-api-access-f8jcm\") pod \"placement-operator-controller-manager-5b797b8dff-dvc87\" (UID: \"64f7e2cc-d596-4270-863c-0fa67797d789\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.668804 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs68g\" (UniqueName: \"kubernetes.io/projected/6f6767fe-2e4b-4b7a-b92c-8007220e1f3e-kube-api-access-fs68g\") pod \"ovn-operator-controller-manager-54fc5f65b7-xblqz\" (UID: \"6f6767fe-2e4b-4b7a-b92c-8007220e1f3e\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.670401 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.699122 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmm4m\" (UniqueName: \"kubernetes.io/projected/42f09610-472c-44b1-85d4-55fa62d2c2c1-kube-api-access-fmm4m\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf\" (UID: \"42f09610-472c-44b1-85d4-55fa62d2c2c1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.726514 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stx4d\" (UniqueName: \"kubernetes.io/projected/b4532ccb-d13d-4cd8-9ccb-240cb2d84195-kube-api-access-stx4d\") pod \"telemetry-operator-controller-manager-6d4bf84b58-2h8kk\" (UID: \"b4532ccb-d13d-4cd8-9ccb-240cb2d84195\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.726716 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gvjs\" (UniqueName: \"kubernetes.io/projected/fed38002-b5d5-4bc0-9fde-5cb9e04c5542-kube-api-access-2gvjs\") pod \"swift-operator-controller-manager-d656998f4-dlfx8\" (UID: \"fed38002-b5d5-4bc0-9fde-5cb9e04c5542\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.726744 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpzkd\" (UniqueName: \"kubernetes.io/projected/82fc1284-da3b-4620-ae81-64535b5dde90-kube-api-access-tpzkd\") pod \"watcher-operator-controller-manager-8c6448b9f-54vjr\" (UID: \"82fc1284-da3b-4620-ae81-64535b5dde90\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.726792 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpcgd\" (UniqueName: \"kubernetes.io/projected/63a16eba-7d99-41f4-9c77-83a8497c4210-kube-api-access-tpcgd\") pod \"test-operator-controller-manager-b4c496f69-j5gtk\" (UID: \"63a16eba-7d99-41f4-9c77-83a8497c4210\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.739417 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.754618 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stx4d\" (UniqueName: \"kubernetes.io/projected/b4532ccb-d13d-4cd8-9ccb-240cb2d84195-kube-api-access-stx4d\") pod \"telemetry-operator-controller-manager-6d4bf84b58-2h8kk\" (UID: \"b4532ccb-d13d-4cd8-9ccb-240cb2d84195\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.763536 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.765535 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gvjs\" (UniqueName: \"kubernetes.io/projected/fed38002-b5d5-4bc0-9fde-5cb9e04c5542-kube-api-access-2gvjs\") pod \"swift-operator-controller-manager-d656998f4-dlfx8\" (UID: \"fed38002-b5d5-4bc0-9fde-5cb9e04c5542\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.783238 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.784342 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.787270 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpcgd\" (UniqueName: \"kubernetes.io/projected/63a16eba-7d99-41f4-9c77-83a8497c4210-kube-api-access-tpcgd\") pod \"test-operator-controller-manager-b4c496f69-j5gtk\" (UID: \"63a16eba-7d99-41f4-9c77-83a8497c4210\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.799621 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpzkd\" (UniqueName: \"kubernetes.io/projected/82fc1284-da3b-4620-ae81-64535b5dde90-kube-api-access-tpzkd\") pod \"watcher-operator-controller-manager-8c6448b9f-54vjr\" (UID: \"82fc1284-da3b-4620-ae81-64535b5dde90\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.804921 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.805668 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-2vs7j" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.819426 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.828918 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.837858 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26e6c9ac-040e-4e10-97ca-6000969950f3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-gvzfc\" (UID: \"26e6c9ac-040e-4e10-97ca-6000969950f3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.850887 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26e6c9ac-040e-4e10-97ca-6000969950f3-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-gvzfc\" (UID: \"26e6c9ac-040e-4e10-97ca-6000969950f3\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.885721 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.913268 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.934622 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.938163 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.946325 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.952256 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z"] Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.952374 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.959221 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-cqhgd" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.990464 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f07b6604-fc2f-4507-a52c-c69c58fc4cee-cert\") pod \"openstack-operator-controller-manager-5f88c7d9f9-wwjhr\" (UID: \"f07b6604-fc2f-4507-a52c-c69c58fc4cee\") " pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" Nov 24 09:08:19 crc kubenswrapper[4944]: I1124 09:08:19.995286 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6zsk\" (UniqueName: \"kubernetes.io/projected/f07b6604-fc2f-4507-a52c-c69c58fc4cee-kube-api-access-s6zsk\") pod \"openstack-operator-controller-manager-5f88c7d9f9-wwjhr\" (UID: \"f07b6604-fc2f-4507-a52c-c69c58fc4cee\") " pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.101622 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6zsk\" (UniqueName: \"kubernetes.io/projected/f07b6604-fc2f-4507-a52c-c69c58fc4cee-kube-api-access-s6zsk\") pod \"openstack-operator-controller-manager-5f88c7d9f9-wwjhr\" (UID: \"f07b6604-fc2f-4507-a52c-c69c58fc4cee\") " pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.101864 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f07b6604-fc2f-4507-a52c-c69c58fc4cee-cert\") pod \"openstack-operator-controller-manager-5f88c7d9f9-wwjhr\" (UID: \"f07b6604-fc2f-4507-a52c-c69c58fc4cee\") " pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.101959 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhk2m\" (UniqueName: \"kubernetes.io/projected/18a65aee-1bdc-4d96-8c52-711fb8cf5476-kube-api-access-xhk2m\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z\" (UID: \"18a65aee-1bdc-4d96-8c52-711fb8cf5476\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z" Nov 24 09:08:20 crc kubenswrapper[4944]: E1124 09:08:20.102826 4944 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 09:08:20 crc kubenswrapper[4944]: E1124 09:08:20.102932 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f07b6604-fc2f-4507-a52c-c69c58fc4cee-cert podName:f07b6604-fc2f-4507-a52c-c69c58fc4cee nodeName:}" failed. No retries permitted until 2025-11-24 09:08:20.60288565 +0000 UTC m=+961.137326112 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f07b6604-fc2f-4507-a52c-c69c58fc4cee-cert") pod "openstack-operator-controller-manager-5f88c7d9f9-wwjhr" (UID: "f07b6604-fc2f-4507-a52c-c69c58fc4cee") : secret "webhook-server-cert" not found Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.135905 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6zsk\" (UniqueName: \"kubernetes.io/projected/f07b6604-fc2f-4507-a52c-c69c58fc4cee-kube-api-access-s6zsk\") pod \"openstack-operator-controller-manager-5f88c7d9f9-wwjhr\" (UID: \"f07b6604-fc2f-4507-a52c-c69c58fc4cee\") " pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.184152 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g"] Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.203323 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhk2m\" (UniqueName: \"kubernetes.io/projected/18a65aee-1bdc-4d96-8c52-711fb8cf5476-kube-api-access-xhk2m\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z\" (UID: \"18a65aee-1bdc-4d96-8c52-711fb8cf5476\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z" Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.203389 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42f09610-472c-44b1-85d4-55fa62d2c2c1-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf\" (UID: \"42f09610-472c-44b1-85d4-55fa62d2c2c1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:20 crc kubenswrapper[4944]: E1124 09:08:20.203743 4944 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 09:08:20 crc kubenswrapper[4944]: E1124 09:08:20.203792 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/42f09610-472c-44b1-85d4-55fa62d2c2c1-cert podName:42f09610-472c-44b1-85d4-55fa62d2c2c1 nodeName:}" failed. No retries permitted until 2025-11-24 09:08:21.20377581 +0000 UTC m=+961.738216272 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/42f09610-472c-44b1-85d4-55fa62d2c2c1-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" (UID: "42f09610-472c-44b1-85d4-55fa62d2c2c1") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.245202 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhk2m\" (UniqueName: \"kubernetes.io/projected/18a65aee-1bdc-4d96-8c52-711fb8cf5476-kube-api-access-xhk2m\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z\" (UID: \"18a65aee-1bdc-4d96-8c52-711fb8cf5476\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z" Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.298271 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-cqhgd" Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.299246 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z" Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.314837 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7"] Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.614069 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f07b6604-fc2f-4507-a52c-c69c58fc4cee-cert\") pod \"openstack-operator-controller-manager-5f88c7d9f9-wwjhr\" (UID: \"f07b6604-fc2f-4507-a52c-c69c58fc4cee\") " pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.621403 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7" event={"ID":"f2b53031-1259-46ff-a42d-f65547877116","Type":"ContainerStarted","Data":"2efb597797c5347914bbd476dbfefd954de97adb5b584e9745d8238931116dea"} Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.621843 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f07b6604-fc2f-4507-a52c-c69c58fc4cee-cert\") pod \"openstack-operator-controller-manager-5f88c7d9f9-wwjhr\" (UID: \"f07b6604-fc2f-4507-a52c-c69c58fc4cee\") " pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.636812 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g" event={"ID":"2640bc37-2c54-4e38-ae79-a1d5cc08af65","Type":"ContainerStarted","Data":"c11bb754c905d834e19de712a46fe824ae651c474e8e8bc479adb47b2ba328fe"} Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.722884 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg"] Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.731000 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn"] Nov 24 09:08:20 crc kubenswrapper[4944]: W1124 09:08:20.733934 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28277a17_6dab_4f8e_b765_46fefba90b01.slice/crio-dbe093cfd3224359b1c7ef76db5f4a24e48e71d651242b6169b44c2420fd8123 WatchSource:0}: Error finding container dbe093cfd3224359b1c7ef76db5f4a24e48e71d651242b6169b44c2420fd8123: Status 404 returned error can't find the container with id dbe093cfd3224359b1c7ef76db5f4a24e48e71d651242b6169b44c2420fd8123 Nov 24 09:08:20 crc kubenswrapper[4944]: W1124 09:08:20.738156 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod916de258_fc43_4fcc_bd66_9e24e3216ac8.slice/crio-529b7ea191b3d8702fcb2f9995e6fe04fb422141d422ed6875e786758ef44653 WatchSource:0}: Error finding container 529b7ea191b3d8702fcb2f9995e6fe04fb422141d422ed6875e786758ef44653: Status 404 returned error can't find the container with id 529b7ea191b3d8702fcb2f9995e6fe04fb422141d422ed6875e786758ef44653 Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.879641 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-2vs7j" Nov 24 09:08:20 crc kubenswrapper[4944]: I1124 09:08:20.887987 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.142688 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.152638 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.193295 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6"] Nov 24 09:08:21 crc kubenswrapper[4944]: W1124 09:08:21.195229 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod163609af_d6a1_48b2_a5cb_e844edff0ef4.slice/crio-2a4155a11193034acd3651438572387ac4f64a0cad1058c32e557ee18543cecf WatchSource:0}: Error finding container 2a4155a11193034acd3651438572387ac4f64a0cad1058c32e557ee18543cecf: Status 404 returned error can't find the container with id 2a4155a11193034acd3651438572387ac4f64a0cad1058c32e557ee18543cecf Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.198923 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.230402 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42f09610-472c-44b1-85d4-55fa62d2c2c1-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf\" (UID: \"42f09610-472c-44b1-85d4-55fa62d2c2c1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.241787 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42f09610-472c-44b1-85d4-55fa62d2c2c1-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf\" (UID: \"42f09610-472c-44b1-85d4-55fa62d2c2c1\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.242399 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r"] Nov 24 09:08:21 crc kubenswrapper[4944]: W1124 09:08:21.244648 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e0e56a5_c24a_4dea_b478_d742409c80b4.slice/crio-3ff6c692d3fb4cd8568050cb74080ff0d6a83e46ef3e41d793f578b0c51a9cb4 WatchSource:0}: Error finding container 3ff6c692d3fb4cd8568050cb74080ff0d6a83e46ef3e41d793f578b0c51a9cb4: Status 404 returned error can't find the container with id 3ff6c692d3fb4cd8568050cb74080ff0d6a83e46ef3e41d793f578b0c51a9cb4 Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.258715 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.270494 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.429838 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.443671 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.475860 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.485826 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.492031 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.507059 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z"] Nov 24 09:08:21 crc kubenswrapper[4944]: W1124 09:08:21.514882 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc702883d_7a93_4ede_a1dc_b2becaf94999.slice/crio-acbe3388bdf499c84f5024a44f91731a9fd7e424cde5602c0b65026479b4b865 WatchSource:0}: Error finding container acbe3388bdf499c84f5024a44f91731a9fd7e424cde5602c0b65026479b4b865: Status 404 returned error can't find the container with id acbe3388bdf499c84f5024a44f91731a9fd7e424cde5602c0b65026479b4b865 Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.514954 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.516690 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xswz5" Nov 24 09:08:21 crc kubenswrapper[4944]: W1124 09:08:21.517393 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26e6c9ac_040e_4e10_97ca_6000969950f3.slice/crio-604ac6901cc9c6734179c67817a570eda80b9b485f7e8be1ad25f0252ab3dd8b WatchSource:0}: Error finding container 604ac6901cc9c6734179c67817a570eda80b9b485f7e8be1ad25f0252ab3dd8b: Status 404 returned error can't find the container with id 604ac6901cc9c6734179c67817a570eda80b9b485f7e8be1ad25f0252ab3dd8b Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.522413 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tpcgd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-j5gtk_openstack-operators(63a16eba-7d99-41f4-9c77-83a8497c4210): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.523903 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tpzkd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-54vjr_openstack-operators(82fc1284-da3b-4620-ae81-64535b5dde90): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.524825 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.527970 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xhk2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z_openstack-operators(18a65aee-1bdc-4d96-8c52-711fb8cf5476): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.529176 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z" podUID="18a65aee-1bdc-4d96-8c52-711fb8cf5476" Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.530153 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v6qtr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-54cfbf4c7d-gcggn_openstack-operators(c702883d-7a93-4ede-a1dc-b2becaf94999): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 09:08:21 crc kubenswrapper[4944]: W1124 09:08:21.533362 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf62b2a83_584b_4f8a_aa62_c1e7499100bd.slice/crio-975a8dacac2648943ecfe2a4f203f156d436541ffb7735ca2e8dfc832b50868b WatchSource:0}: Error finding container 975a8dacac2648943ecfe2a4f203f156d436541ffb7735ca2e8dfc832b50868b: Status 404 returned error can't find the container with id 975a8dacac2648943ecfe2a4f203f156d436541ffb7735ca2e8dfc832b50868b Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.539346 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.546458 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr"] Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.557829 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kdpl9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-58f887965d-d5mp6_openstack-operators(f62b2a83-584b-4f8a-aa62-c1e7499100bd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.557953 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2gvjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d656998f4-dlfx8_openstack-operators(fed38002-b5d5-4bc0-9fde-5cb9e04c5542): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.565016 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.574437 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr"] Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.646344 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl" event={"ID":"163609af-d6a1-48b2-a5cb-e844edff0ef4","Type":"ContainerStarted","Data":"2a4155a11193034acd3651438572387ac4f64a0cad1058c32e557ee18543cecf"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.651847 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn" event={"ID":"916de258-fc43-4fcc-bd66-9e24e3216ac8","Type":"ContainerStarted","Data":"529b7ea191b3d8702fcb2f9995e6fe04fb422141d422ed6875e786758ef44653"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.654234 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" event={"ID":"26e6c9ac-040e-4e10-97ca-6000969950f3","Type":"ContainerStarted","Data":"604ac6901cc9c6734179c67817a570eda80b9b485f7e8be1ad25f0252ab3dd8b"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.656446 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6" event={"ID":"f5f1851e-a78f-458f-9c7a-fa271cfa4283","Type":"ContainerStarted","Data":"e36e0216c5334bdc7e5db04abbe5231a47bf7b7b3bef462b69ee1c51115f21c7"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.658536 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" event={"ID":"63a16eba-7d99-41f4-9c77-83a8497c4210","Type":"ContainerStarted","Data":"efbbca91ce2826c9d76004bfb8be97591416b26f13defe953c3f77bf830d23fa"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.659818 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" event={"ID":"82fc1284-da3b-4620-ae81-64535b5dde90","Type":"ContainerStarted","Data":"e900318dc2fe77e3d4535c19d9c1ca8c54f43929587ef17903a606b4fbe74f86"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.661959 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" event={"ID":"f07b6604-fc2f-4507-a52c-c69c58fc4cee","Type":"ContainerStarted","Data":"a63b99bfb52ffdac3b3a1ebac8a0c54f226e7be1e524aac2b42384d8efc8af3e"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.664078 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk" event={"ID":"b4532ccb-d13d-4cd8-9ccb-240cb2d84195","Type":"ContainerStarted","Data":"768509b6cbf5cf1768c73e5e2aeaf47583c6bf2bb39095655740f021ce8022d3"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.665308 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87" event={"ID":"64f7e2cc-d596-4270-863c-0fa67797d789","Type":"ContainerStarted","Data":"d8c9270078155a5bc52fa3d0c1234bb478d88bf40eed0d884d288eb93c040ef9"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.666701 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk" event={"ID":"655196aa-961a-4b69-8fad-6975c0ce44f2","Type":"ContainerStarted","Data":"607f431f9295a4ce48cef4e81c724389e327d73421eecf015daddca672b2357e"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.668282 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z" event={"ID":"18a65aee-1bdc-4d96-8c52-711fb8cf5476","Type":"ContainerStarted","Data":"4a468582c3476105c4df5fa632b3fa2bc1a97d26d7e02573a7ca66b3e2bcc43e"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.670409 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" event={"ID":"fed38002-b5d5-4bc0-9fde-5cb9e04c5542","Type":"ContainerStarted","Data":"8d64d434a2035ff8bc091b5a6712d4e3f1dd74144f8c58e48ba575f6b579f36f"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.671989 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k" event={"ID":"265e80c3-fc9b-43f0-a8f8-6f3a4562a9ca","Type":"ContainerStarted","Data":"a3b2094f1f1697d757585e98bb958a57e8547d25c106994a446c7f8c58630d8b"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.673222 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" event={"ID":"f62b2a83-584b-4f8a-aa62-c1e7499100bd","Type":"ContainerStarted","Data":"975a8dacac2648943ecfe2a4f203f156d436541ffb7735ca2e8dfc832b50868b"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.674395 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz" event={"ID":"fe8abc9f-e87f-4417-905f-44bb0b432a59","Type":"ContainerStarted","Data":"017728ed434b5aa2fb94e37e49cd222408bc7ff708f8348ef3b62a105cce7e84"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.676197 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg" event={"ID":"28277a17-6dab-4f8e-b765-46fefba90b01","Type":"ContainerStarted","Data":"dbe093cfd3224359b1c7ef76db5f4a24e48e71d651242b6169b44c2420fd8123"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.677353 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz" event={"ID":"6f6767fe-2e4b-4b7a-b92c-8007220e1f3e","Type":"ContainerStarted","Data":"67b1dda6a02f692a6c804005088c79fbd06256bf2e6b6b15704d257a7d88b430"} Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.692217 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z" podUID="18a65aee-1bdc-4d96-8c52-711fb8cf5476" Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.696274 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r" event={"ID":"ec7ed47f-2c2c-4401-ac0a-0510d8649f01","Type":"ContainerStarted","Data":"0ea14b839ab695c737898158a2818aeb874aa5be28322ec0dda6a79bb6dbe788"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.699319 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" event={"ID":"c702883d-7a93-4ede-a1dc-b2becaf94999","Type":"ContainerStarted","Data":"acbe3388bdf499c84f5024a44f91731a9fd7e424cde5602c0b65026479b4b865"} Nov 24 09:08:21 crc kubenswrapper[4944]: I1124 09:08:21.700739 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp" event={"ID":"3e0e56a5-c24a-4dea-b478-d742409c80b4","Type":"ContainerStarted","Data":"3ff6c692d3fb4cd8568050cb74080ff0d6a83e46ef3e41d793f578b0c51a9cb4"} Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.870935 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" podUID="63a16eba-7d99-41f4-9c77-83a8497c4210" Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.949572 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" podUID="82fc1284-da3b-4620-ae81-64535b5dde90" Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.972143 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" podUID="f62b2a83-584b-4f8a-aa62-c1e7499100bd" Nov 24 09:08:21 crc kubenswrapper[4944]: E1124 09:08:21.984250 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" podUID="c702883d-7a93-4ede-a1dc-b2becaf94999" Nov 24 09:08:22 crc kubenswrapper[4944]: E1124 09:08:22.033689 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" podUID="fed38002-b5d5-4bc0-9fde-5cb9e04c5542" Nov 24 09:08:22 crc kubenswrapper[4944]: I1124 09:08:22.152266 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf"] Nov 24 09:08:22 crc kubenswrapper[4944]: I1124 09:08:22.740637 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" event={"ID":"f62b2a83-584b-4f8a-aa62-c1e7499100bd","Type":"ContainerStarted","Data":"18ff6a2153e58630542342388d5d0603bd18148e3bf0f79f6b0de9208010483b"} Nov 24 09:08:22 crc kubenswrapper[4944]: I1124 09:08:22.744930 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" event={"ID":"fed38002-b5d5-4bc0-9fde-5cb9e04c5542","Type":"ContainerStarted","Data":"57f2470486936cf78d8e0cf57d0b3e3404daa14d4ec94af46b5e4e62c82a41d9"} Nov 24 09:08:22 crc kubenswrapper[4944]: E1124 09:08:22.746654 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a\\\"\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" podUID="f62b2a83-584b-4f8a-aa62-c1e7499100bd" Nov 24 09:08:22 crc kubenswrapper[4944]: E1124 09:08:22.747194 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" podUID="fed38002-b5d5-4bc0-9fde-5cb9e04c5542" Nov 24 09:08:22 crc kubenswrapper[4944]: I1124 09:08:22.759418 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" event={"ID":"63a16eba-7d99-41f4-9c77-83a8497c4210","Type":"ContainerStarted","Data":"d38148092aacbfed7996a738c71ef180619c13677bc09d5f8305967919dc3904"} Nov 24 09:08:22 crc kubenswrapper[4944]: E1124 09:08:22.761407 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" podUID="63a16eba-7d99-41f4-9c77-83a8497c4210" Nov 24 09:08:22 crc kubenswrapper[4944]: I1124 09:08:22.767099 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" event={"ID":"82fc1284-da3b-4620-ae81-64535b5dde90","Type":"ContainerStarted","Data":"3d6d082ad321564229fd75f5b5bf05ebdedde04381c5adc7d755aded91249954"} Nov 24 09:08:22 crc kubenswrapper[4944]: E1124 09:08:22.768602 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" podUID="82fc1284-da3b-4620-ae81-64535b5dde90" Nov 24 09:08:22 crc kubenswrapper[4944]: I1124 09:08:22.779732 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" event={"ID":"f07b6604-fc2f-4507-a52c-c69c58fc4cee","Type":"ContainerStarted","Data":"0578a8ca30f9a930bb88e42c6f32ddead15e429b6f9d4f8b3b8fe2bd00f18c7a"} Nov 24 09:08:22 crc kubenswrapper[4944]: I1124 09:08:22.779774 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" event={"ID":"f07b6604-fc2f-4507-a52c-c69c58fc4cee","Type":"ContainerStarted","Data":"8f5e8adad57fec89eeef3d1ffe3a967a385728ee6bc99fcacfb7b779a39cea93"} Nov 24 09:08:22 crc kubenswrapper[4944]: I1124 09:08:22.780547 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" Nov 24 09:08:22 crc kubenswrapper[4944]: I1124 09:08:22.801937 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" event={"ID":"c702883d-7a93-4ede-a1dc-b2becaf94999","Type":"ContainerStarted","Data":"7a5d262295425118c4071e2b6281723c4b478493e4cb6cc205f88b96d2daca21"} Nov 24 09:08:22 crc kubenswrapper[4944]: E1124 09:08:22.803622 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" podUID="c702883d-7a93-4ede-a1dc-b2becaf94999" Nov 24 09:08:22 crc kubenswrapper[4944]: I1124 09:08:22.821148 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" event={"ID":"42f09610-472c-44b1-85d4-55fa62d2c2c1","Type":"ContainerStarted","Data":"679af736be333925c8b0229d4ea305aa08e193dfe9daa870e8a6023bec21b73e"} Nov 24 09:08:22 crc kubenswrapper[4944]: E1124 09:08:22.823848 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z" podUID="18a65aee-1bdc-4d96-8c52-711fb8cf5476" Nov 24 09:08:22 crc kubenswrapper[4944]: I1124 09:08:22.963543 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" podStartSLOduration=3.963513612 podStartE2EDuration="3.963513612s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:08:22.951282429 +0000 UTC m=+963.485722912" watchObservedRunningTime="2025-11-24 09:08:22.963513612 +0000 UTC m=+963.497954074" Nov 24 09:08:23 crc kubenswrapper[4944]: E1124 09:08:23.831143 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" podUID="c702883d-7a93-4ede-a1dc-b2becaf94999" Nov 24 09:08:23 crc kubenswrapper[4944]: E1124 09:08:23.831339 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" podUID="63a16eba-7d99-41f4-9c77-83a8497c4210" Nov 24 09:08:23 crc kubenswrapper[4944]: E1124 09:08:23.831552 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" podUID="fed38002-b5d5-4bc0-9fde-5cb9e04c5542" Nov 24 09:08:23 crc kubenswrapper[4944]: E1124 09:08:23.831597 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" podUID="82fc1284-da3b-4620-ae81-64535b5dde90" Nov 24 09:08:23 crc kubenswrapper[4944]: E1124 09:08:23.831737 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a\\\"\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" podUID="f62b2a83-584b-4f8a-aa62-c1e7499100bd" Nov 24 09:08:30 crc kubenswrapper[4944]: I1124 09:08:30.901458 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5f88c7d9f9-wwjhr" Nov 24 09:08:34 crc kubenswrapper[4944]: I1124 09:08:34.996367 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn" event={"ID":"916de258-fc43-4fcc-bd66-9e24e3216ac8","Type":"ContainerStarted","Data":"437994f6a63e06879d8cf3400ba842166a86259aa0b3408005ee52af3c18b640"} Nov 24 09:08:34 crc kubenswrapper[4944]: I1124 09:08:34.996915 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn" event={"ID":"916de258-fc43-4fcc-bd66-9e24e3216ac8","Type":"ContainerStarted","Data":"0a241794f09e3684d2b3ff544b5180f5e0de9b815cbabddf6ac091ff6ff8ce16"} Nov 24 09:08:34 crc kubenswrapper[4944]: I1124 09:08:34.996962 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn" Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.045767 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" event={"ID":"26e6c9ac-040e-4e10-97ca-6000969950f3","Type":"ContainerStarted","Data":"990582c5f756d21fdbd28b1ffb3cc24ea5c2b0539b900c866ee240b4988b441f"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.056750 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn" podStartSLOduration=4.092747647 podStartE2EDuration="17.056726424s" podCreationTimestamp="2025-11-24 09:08:18 +0000 UTC" firstStartedPulling="2025-11-24 09:08:20.741915033 +0000 UTC m=+961.276355495" lastFinishedPulling="2025-11-24 09:08:33.70589381 +0000 UTC m=+974.240334272" observedRunningTime="2025-11-24 09:08:35.032814551 +0000 UTC m=+975.567255013" watchObservedRunningTime="2025-11-24 09:08:35.056726424 +0000 UTC m=+975.591166896" Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.070916 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk" event={"ID":"b4532ccb-d13d-4cd8-9ccb-240cb2d84195","Type":"ContainerStarted","Data":"3cca458a58f8ecc746471e735ea6da6680d9875277096fbe240bbe6a10066002"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.072084 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk" Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.096745 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87" event={"ID":"64f7e2cc-d596-4270-863c-0fa67797d789","Type":"ContainerStarted","Data":"a76edce0d6af27a99666e76d123591ce1d8614a8a0087eccf9618d5f26b2db21"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.126254 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk" podStartSLOduration=3.859118586 podStartE2EDuration="16.126238059s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.453460945 +0000 UTC m=+961.987901407" lastFinishedPulling="2025-11-24 09:08:33.720580418 +0000 UTC m=+974.255020880" observedRunningTime="2025-11-24 09:08:35.123308625 +0000 UTC m=+975.657749107" watchObservedRunningTime="2025-11-24 09:08:35.126238059 +0000 UTC m=+975.660678511" Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.126420 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz" event={"ID":"6f6767fe-2e4b-4b7a-b92c-8007220e1f3e","Type":"ContainerStarted","Data":"b9caaf394cfaad3491e18df9965833afffe66a2971547a6d007c21ba79c0db37"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.160660 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz" event={"ID":"fe8abc9f-e87f-4417-905f-44bb0b432a59","Type":"ContainerStarted","Data":"037482421ca3d18fa27911512d4f3ac66278fd0740eed246b18ac9b87f444be6"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.160711 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz" event={"ID":"fe8abc9f-e87f-4417-905f-44bb0b432a59","Type":"ContainerStarted","Data":"2e41798946c736749cb7a4cc7af3441c3329d96b352ea30d20d795bd4ccff816"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.161192 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz" Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.180841 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg" event={"ID":"28277a17-6dab-4f8e-b765-46fefba90b01","Type":"ContainerStarted","Data":"68c2e299aec45a81a0ae60d09eb67d303ccd483973faedc8e317980b95e1ec4c"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.198106 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r" event={"ID":"ec7ed47f-2c2c-4401-ac0a-0510d8649f01","Type":"ContainerStarted","Data":"6ea92417ad70103099c99f231b8c3ca298ba065c8aea93cebf50750b1d07e574"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.200267 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz" podStartSLOduration=3.751003944 podStartE2EDuration="16.200237328s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.301806705 +0000 UTC m=+961.836247167" lastFinishedPulling="2025-11-24 09:08:33.751040089 +0000 UTC m=+974.285480551" observedRunningTime="2025-11-24 09:08:35.18839874 +0000 UTC m=+975.722839222" watchObservedRunningTime="2025-11-24 09:08:35.200237328 +0000 UTC m=+975.734677790" Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.229723 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k" event={"ID":"265e80c3-fc9b-43f0-a8f8-6f3a4562a9ca","Type":"ContainerStarted","Data":"4b19d31d7b8ac990e2960e759f8cc60d1f8b74390265fdac7e9f7eeb5487dcab"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.231350 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" event={"ID":"42f09610-472c-44b1-85d4-55fa62d2c2c1","Type":"ContainerStarted","Data":"3e57de0ba68d31f3cd7ac1605c8090e075edd123853c93b3fc3ce52b509a3aed"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.232571 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk" event={"ID":"655196aa-961a-4b69-8fad-6975c0ce44f2","Type":"ContainerStarted","Data":"97cd185026e726d5979455e6255b52d575c6cd26d1d580e0c8813db9c4e1bdd4"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.233655 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6" event={"ID":"f5f1851e-a78f-458f-9c7a-fa271cfa4283","Type":"ContainerStarted","Data":"007f093a3b26bcb3891736b79c677687ca557789332499acd73e962f414938c6"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.247660 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g" event={"ID":"2640bc37-2c54-4e38-ae79-a1d5cc08af65","Type":"ContainerStarted","Data":"5258215d01576d2675d56166290ddc24b28dc3fde113a9ad66d0d22299065c8f"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.247708 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g" event={"ID":"2640bc37-2c54-4e38-ae79-a1d5cc08af65","Type":"ContainerStarted","Data":"31fccc7475f66c87cb98460de9c0741120278a4a20bb4adc4fd75a6052c821b5"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.248203 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g" Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.265901 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7" event={"ID":"f2b53031-1259-46ff-a42d-f65547877116","Type":"ContainerStarted","Data":"39b6d6555169e5a7db48bfc8fa2d6f602615745e3607a85af5010da6639b514d"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.268415 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl" event={"ID":"163609af-d6a1-48b2-a5cb-e844edff0ef4","Type":"ContainerStarted","Data":"f62b7985a82ba916fb8822692df40fcf3495f897a150cc80383f6fd375c06ac9"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.270635 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp" event={"ID":"3e0e56a5-c24a-4dea-b478-d742409c80b4","Type":"ContainerStarted","Data":"9fdc54e134b8e697bd7630106ec40bbc4814ec930738b771dff2518d39cf8362"} Nov 24 09:08:35 crc kubenswrapper[4944]: I1124 09:08:35.273038 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g" podStartSLOduration=3.92864709 podStartE2EDuration="17.273020087s" podCreationTimestamp="2025-11-24 09:08:18 +0000 UTC" firstStartedPulling="2025-11-24 09:08:20.345476622 +0000 UTC m=+960.879917084" lastFinishedPulling="2025-11-24 09:08:33.689849619 +0000 UTC m=+974.224290081" observedRunningTime="2025-11-24 09:08:35.27216228 +0000 UTC m=+975.806602742" watchObservedRunningTime="2025-11-24 09:08:35.273020087 +0000 UTC m=+975.807460549" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.278983 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.287255 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.287293 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl" event={"ID":"163609af-d6a1-48b2-a5cb-e844edff0ef4","Type":"ContainerStarted","Data":"efa4cf86ed9cd1c23aba764541bbd4eb301314cfe35687ff86f0f23c58e3c3c9"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.288527 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7" event={"ID":"f2b53031-1259-46ff-a42d-f65547877116","Type":"ContainerStarted","Data":"d7d781085d3cbaa96c853b716b7a994d5e41b4dbabc7e0bda430c0ebafdf902a"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.288638 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.291999 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" event={"ID":"42f09610-472c-44b1-85d4-55fa62d2c2c1","Type":"ContainerStarted","Data":"5c45c291b37d274e53078ed87593f8a7f855726e094a25a57cd65febab205988"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.292093 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.309001 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk" event={"ID":"b4532ccb-d13d-4cd8-9ccb-240cb2d84195","Type":"ContainerStarted","Data":"1605724db171c7ae91f913f52bd479c75a002d34ae04c09c1e1e8d234619785c"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.328209 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7" podStartSLOduration=5.045663453 podStartE2EDuration="18.328191827s" podCreationTimestamp="2025-11-24 09:08:18 +0000 UTC" firstStartedPulling="2025-11-24 09:08:20.437328211 +0000 UTC m=+960.971768673" lastFinishedPulling="2025-11-24 09:08:33.719856585 +0000 UTC m=+974.254297047" observedRunningTime="2025-11-24 09:08:36.32703598 +0000 UTC m=+976.861476442" watchObservedRunningTime="2025-11-24 09:08:36.328191827 +0000 UTC m=+976.862632309" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.332837 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl" podStartSLOduration=4.810879125 podStartE2EDuration="17.332814154s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.229395759 +0000 UTC m=+961.763836231" lastFinishedPulling="2025-11-24 09:08:33.751330798 +0000 UTC m=+974.285771260" observedRunningTime="2025-11-24 09:08:36.308466678 +0000 UTC m=+976.842907140" watchObservedRunningTime="2025-11-24 09:08:36.332814154 +0000 UTC m=+976.867254616" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.333821 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87" event={"ID":"64f7e2cc-d596-4270-863c-0fa67797d789","Type":"ContainerStarted","Data":"55e1b330406a8c26a2b367cc12f005efb63a92aa6c60130fd6e5c06aa58910c1"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.334286 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.344698 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg" event={"ID":"28277a17-6dab-4f8e-b765-46fefba90b01","Type":"ContainerStarted","Data":"300492af731a5387f58fe66cca1e58d7a6acb6942fcf4a853c27f130a9cdcc8b"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.355090 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk" event={"ID":"655196aa-961a-4b69-8fad-6975c0ce44f2","Type":"ContainerStarted","Data":"4802ff3e9622abea2e8d3673b1240c1a2f08a6f55e6babd2a2081d4c6156fe58"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.355154 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.377908 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz" event={"ID":"6f6767fe-2e4b-4b7a-b92c-8007220e1f3e","Type":"ContainerStarted","Data":"a63ee9235f2103b6d453ab2144a89ccdeb7f83444d0f99a522bd82c98e94bdf9"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.378259 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.387883 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6" event={"ID":"f5f1851e-a78f-458f-9c7a-fa271cfa4283","Type":"ContainerStarted","Data":"1897ce4ed78bb36a22aec21fcb7f7377cc5e00aeb96b9654c6c22bd0f69dbcb5"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.388075 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.397900 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" podStartSLOduration=5.837514329 podStartE2EDuration="17.397871538s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:22.18565889 +0000 UTC m=+962.720099352" lastFinishedPulling="2025-11-24 09:08:33.746016099 +0000 UTC m=+974.280456561" observedRunningTime="2025-11-24 09:08:36.386252847 +0000 UTC m=+976.920693309" watchObservedRunningTime="2025-11-24 09:08:36.397871538 +0000 UTC m=+976.932312000" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.411364 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r" event={"ID":"ec7ed47f-2c2c-4401-ac0a-0510d8649f01","Type":"ContainerStarted","Data":"fcfea388a2c337fb760e492f2699d4e8a3a8f4686019274233cb2ea800b2d389"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.412277 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.418239 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk" podStartSLOduration=5.941746566 podStartE2EDuration="18.418220976s" podCreationTimestamp="2025-11-24 09:08:18 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.271305795 +0000 UTC m=+961.805746257" lastFinishedPulling="2025-11-24 09:08:33.747780195 +0000 UTC m=+974.282220667" observedRunningTime="2025-11-24 09:08:36.41739185 +0000 UTC m=+976.951832322" watchObservedRunningTime="2025-11-24 09:08:36.418220976 +0000 UTC m=+976.952661458" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.428565 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" event={"ID":"26e6c9ac-040e-4e10-97ca-6000969950f3","Type":"ContainerStarted","Data":"d51153c26669ca15226514efdef692e269812627f004a1a69e0223dc8d7437f3"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.430101 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.433385 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp" event={"ID":"3e0e56a5-c24a-4dea-b478-d742409c80b4","Type":"ContainerStarted","Data":"05bcb9bc6a71fe67980910f2703895630fee0fd19628767978853c17ff241582"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.433515 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.438574 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k" event={"ID":"265e80c3-fc9b-43f0-a8f8-6f3a4562a9ca","Type":"ContainerStarted","Data":"e6973f95d50012187e86e427eaffcb31759225d9c57a6cd505021478b2432e5a"} Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.439036 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.450808 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz" podStartSLOduration=5.219419382 podStartE2EDuration="17.450780234s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.498403009 +0000 UTC m=+962.032843471" lastFinishedPulling="2025-11-24 09:08:33.729763861 +0000 UTC m=+974.264204323" observedRunningTime="2025-11-24 09:08:36.446057464 +0000 UTC m=+976.980497926" watchObservedRunningTime="2025-11-24 09:08:36.450780234 +0000 UTC m=+976.985220686" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.466314 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87" podStartSLOduration=5.235580897 podStartE2EDuration="17.466294498s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.498531103 +0000 UTC m=+962.032971565" lastFinishedPulling="2025-11-24 09:08:33.729244704 +0000 UTC m=+974.263685166" observedRunningTime="2025-11-24 09:08:36.464646706 +0000 UTC m=+976.999087178" watchObservedRunningTime="2025-11-24 09:08:36.466294498 +0000 UTC m=+977.000734960" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.488865 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg" podStartSLOduration=5.505172952 podStartE2EDuration="18.488835547s" podCreationTimestamp="2025-11-24 09:08:18 +0000 UTC" firstStartedPulling="2025-11-24 09:08:20.735915441 +0000 UTC m=+961.270355903" lastFinishedPulling="2025-11-24 09:08:33.719578046 +0000 UTC m=+974.254018498" observedRunningTime="2025-11-24 09:08:36.479154088 +0000 UTC m=+977.013594560" watchObservedRunningTime="2025-11-24 09:08:36.488835547 +0000 UTC m=+977.023276009" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.526547 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" podStartSLOduration=6.32114905 podStartE2EDuration="18.526526788s" podCreationTimestamp="2025-11-24 09:08:18 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.520697385 +0000 UTC m=+962.055137847" lastFinishedPulling="2025-11-24 09:08:33.726075123 +0000 UTC m=+974.260515585" observedRunningTime="2025-11-24 09:08:36.504329801 +0000 UTC m=+977.038770263" watchObservedRunningTime="2025-11-24 09:08:36.526526788 +0000 UTC m=+977.060967250" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.529515 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6" podStartSLOduration=6.060863179 podStartE2EDuration="18.529499493s" podCreationTimestamp="2025-11-24 09:08:18 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.251637234 +0000 UTC m=+961.786077696" lastFinishedPulling="2025-11-24 09:08:33.720273548 +0000 UTC m=+974.254714010" observedRunningTime="2025-11-24 09:08:36.521967572 +0000 UTC m=+977.056408044" watchObservedRunningTime="2025-11-24 09:08:36.529499493 +0000 UTC m=+977.063939955" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.538426 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k" podStartSLOduration=6.016812516 podStartE2EDuration="18.538410417s" podCreationTimestamp="2025-11-24 09:08:18 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.180750077 +0000 UTC m=+961.715190539" lastFinishedPulling="2025-11-24 09:08:33.702347978 +0000 UTC m=+974.236788440" observedRunningTime="2025-11-24 09:08:36.535821064 +0000 UTC m=+977.070261526" watchObservedRunningTime="2025-11-24 09:08:36.538410417 +0000 UTC m=+977.072850869" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.554125 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r" podStartSLOduration=5.097088576 podStartE2EDuration="17.554108807s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.270461848 +0000 UTC m=+961.804902310" lastFinishedPulling="2025-11-24 09:08:33.727482079 +0000 UTC m=+974.261922541" observedRunningTime="2025-11-24 09:08:36.553783307 +0000 UTC m=+977.088223789" watchObservedRunningTime="2025-11-24 09:08:36.554108807 +0000 UTC m=+977.088549269" Nov 24 09:08:36 crc kubenswrapper[4944]: I1124 09:08:36.575274 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp" podStartSLOduration=5.105008084 podStartE2EDuration="17.575250411s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.246688375 +0000 UTC m=+961.781128847" lastFinishedPulling="2025-11-24 09:08:33.716930712 +0000 UTC m=+974.251371174" observedRunningTime="2025-11-24 09:08:36.571679577 +0000 UTC m=+977.106120049" watchObservedRunningTime="2025-11-24 09:08:36.575250411 +0000 UTC m=+977.109690883" Nov 24 09:08:37 crc kubenswrapper[4944]: I1124 09:08:37.448733 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.184636 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-qvqs7" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.194305 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-cpf2g" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.229170 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-864hg" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.267907 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7969689c84-8b2jn" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.318924 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-pq27k" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.319013 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-bnxrk" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.422733 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-jq6w6" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.468900 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" event={"ID":"63a16eba-7d99-41f4-9c77-83a8497c4210","Type":"ContainerStarted","Data":"1ecdf93a168acf9a6307479c0cf7fc979d206120c03b5f3244043cb16037162e"} Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.485618 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" podStartSLOduration=3.598108594 podStartE2EDuration="20.485599768s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.522246104 +0000 UTC m=+962.056686566" lastFinishedPulling="2025-11-24 09:08:38.409737278 +0000 UTC m=+978.944177740" observedRunningTime="2025-11-24 09:08:39.481182857 +0000 UTC m=+980.015623349" watchObservedRunningTime="2025-11-24 09:08:39.485599768 +0000 UTC m=+980.020040230" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.604493 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-wjthz" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.631339 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-td8rp" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.657151 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-p4tcl" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.684591 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-xblqz" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.742569 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-wgm4r" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.772029 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-dvc87" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.858220 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-2h8kk" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.944502 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-gvzfc" Nov 24 09:08:39 crc kubenswrapper[4944]: I1124 09:08:39.955384 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" Nov 24 09:08:40 crc kubenswrapper[4944]: I1124 09:08:40.481772 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" event={"ID":"fed38002-b5d5-4bc0-9fde-5cb9e04c5542","Type":"ContainerStarted","Data":"bef5facf56d376de12749bed2087561d38dc7aa9636799421874727ff5bfd90e"} Nov 24 09:08:40 crc kubenswrapper[4944]: I1124 09:08:40.482492 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" Nov 24 09:08:40 crc kubenswrapper[4944]: I1124 09:08:40.506880 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" podStartSLOduration=4.655545045 podStartE2EDuration="21.506861097s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.557753165 +0000 UTC m=+962.092193627" lastFinishedPulling="2025-11-24 09:08:38.409069217 +0000 UTC m=+978.943509679" observedRunningTime="2025-11-24 09:08:40.500747082 +0000 UTC m=+981.035187544" watchObservedRunningTime="2025-11-24 09:08:40.506861097 +0000 UTC m=+981.041301559" Nov 24 09:08:41 crc kubenswrapper[4944]: I1124 09:08:41.492599 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" event={"ID":"c702883d-7a93-4ede-a1dc-b2becaf94999","Type":"ContainerStarted","Data":"936ec20784eb66de5f6deb9115dfbea94ecf9da125c1288be42de2c9ac83347f"} Nov 24 09:08:41 crc kubenswrapper[4944]: I1124 09:08:41.493172 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" Nov 24 09:08:41 crc kubenswrapper[4944]: I1124 09:08:41.496076 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" event={"ID":"f62b2a83-584b-4f8a-aa62-c1e7499100bd","Type":"ContainerStarted","Data":"8ded1c43d59193631d2f2e9e0719497d6ddb3f591e4217af67fb961bb63f6196"} Nov 24 09:08:41 crc kubenswrapper[4944]: I1124 09:08:41.496295 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" Nov 24 09:08:41 crc kubenswrapper[4944]: I1124 09:08:41.500326 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z" event={"ID":"18a65aee-1bdc-4d96-8c52-711fb8cf5476","Type":"ContainerStarted","Data":"51aa6bbaa5dba8484d019bb4e709a781fedef5c8b867df22b0ac8a7b0c35d50a"} Nov 24 09:08:41 crc kubenswrapper[4944]: I1124 09:08:41.511217 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" podStartSLOduration=3.002951356 podStartE2EDuration="22.511199426s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.530039944 +0000 UTC m=+962.064480406" lastFinishedPulling="2025-11-24 09:08:41.038288014 +0000 UTC m=+981.572728476" observedRunningTime="2025-11-24 09:08:41.511036061 +0000 UTC m=+982.045476543" watchObservedRunningTime="2025-11-24 09:08:41.511199426 +0000 UTC m=+982.045639888" Nov 24 09:08:41 crc kubenswrapper[4944]: I1124 09:08:41.533184 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf" Nov 24 09:08:41 crc kubenswrapper[4944]: I1124 09:08:41.535140 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" podStartSLOduration=3.856767777 podStartE2EDuration="22.53512647s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.557679273 +0000 UTC m=+962.092119735" lastFinishedPulling="2025-11-24 09:08:40.236037966 +0000 UTC m=+980.770478428" observedRunningTime="2025-11-24 09:08:41.533108484 +0000 UTC m=+982.067548956" watchObservedRunningTime="2025-11-24 09:08:41.53512647 +0000 UTC m=+982.069566932" Nov 24 09:08:41 crc kubenswrapper[4944]: I1124 09:08:41.550570 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z" podStartSLOduration=2.975543993 podStartE2EDuration="22.550553021s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.527793832 +0000 UTC m=+962.062234294" lastFinishedPulling="2025-11-24 09:08:41.10280286 +0000 UTC m=+981.637243322" observedRunningTime="2025-11-24 09:08:41.547549515 +0000 UTC m=+982.081989977" watchObservedRunningTime="2025-11-24 09:08:41.550553021 +0000 UTC m=+982.084993483" Nov 24 09:08:43 crc kubenswrapper[4944]: I1124 09:08:43.513726 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" event={"ID":"82fc1284-da3b-4620-ae81-64535b5dde90","Type":"ContainerStarted","Data":"f26971c36ad9188cef662a8278169f1233a1ffd08988be81921e92d65ad60e69"} Nov 24 09:08:43 crc kubenswrapper[4944]: I1124 09:08:43.514595 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" Nov 24 09:08:43 crc kubenswrapper[4944]: I1124 09:08:43.530956 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" podStartSLOduration=3.085838907 podStartE2EDuration="24.530936249s" podCreationTimestamp="2025-11-24 09:08:19 +0000 UTC" firstStartedPulling="2025-11-24 09:08:21.523787804 +0000 UTC m=+962.058228266" lastFinishedPulling="2025-11-24 09:08:42.968885146 +0000 UTC m=+983.503325608" observedRunningTime="2025-11-24 09:08:43.530724952 +0000 UTC m=+984.065165424" watchObservedRunningTime="2025-11-24 09:08:43.530936249 +0000 UTC m=+984.065376711" Nov 24 09:08:49 crc kubenswrapper[4944]: I1124 09:08:49.508390 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58f887965d-d5mp6" Nov 24 09:08:49 crc kubenswrapper[4944]: I1124 09:08:49.660507 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-gcggn" Nov 24 09:08:49 crc kubenswrapper[4944]: I1124 09:08:49.889025 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d656998f4-dlfx8" Nov 24 09:08:49 crc kubenswrapper[4944]: I1124 09:08:49.916760 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-54vjr" Nov 24 09:08:49 crc kubenswrapper[4944]: I1124 09:08:49.951181 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-b4c496f69-j5gtk" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.798977 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fdkn4"] Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.807167 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.810849 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fdkn4"] Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.811779 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-qcmf5" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.812000 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.812153 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.819617 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.840892 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cncj\" (UniqueName: \"kubernetes.io/projected/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-kube-api-access-6cncj\") pod \"dnsmasq-dns-675f4bcbfc-fdkn4\" (UID: \"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.840957 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-config\") pod \"dnsmasq-dns-675f4bcbfc-fdkn4\" (UID: \"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.890642 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-852td"] Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.891952 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.898199 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.908535 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-852td"] Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.942436 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-852td\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.942475 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqhfn\" (UniqueName: \"kubernetes.io/projected/eab6b2a1-815c-473a-ac11-b10194ab0e37-kube-api-access-qqhfn\") pod \"dnsmasq-dns-78dd6ddcc-852td\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.942523 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cncj\" (UniqueName: \"kubernetes.io/projected/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-kube-api-access-6cncj\") pod \"dnsmasq-dns-675f4bcbfc-fdkn4\" (UID: \"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.942546 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-config\") pod \"dnsmasq-dns-78dd6ddcc-852td\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.942568 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-config\") pod \"dnsmasq-dns-675f4bcbfc-fdkn4\" (UID: \"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.943637 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-config\") pod \"dnsmasq-dns-675f4bcbfc-fdkn4\" (UID: \"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" Nov 24 09:09:05 crc kubenswrapper[4944]: I1124 09:09:05.962558 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cncj\" (UniqueName: \"kubernetes.io/projected/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-kube-api-access-6cncj\") pod \"dnsmasq-dns-675f4bcbfc-fdkn4\" (UID: \"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" Nov 24 09:09:06 crc kubenswrapper[4944]: I1124 09:09:06.043902 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-config\") pod \"dnsmasq-dns-78dd6ddcc-852td\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:06 crc kubenswrapper[4944]: I1124 09:09:06.044686 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-config\") pod \"dnsmasq-dns-78dd6ddcc-852td\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:06 crc kubenswrapper[4944]: I1124 09:09:06.044797 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqhfn\" (UniqueName: \"kubernetes.io/projected/eab6b2a1-815c-473a-ac11-b10194ab0e37-kube-api-access-qqhfn\") pod \"dnsmasq-dns-78dd6ddcc-852td\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:06 crc kubenswrapper[4944]: I1124 09:09:06.044822 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-852td\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:06 crc kubenswrapper[4944]: I1124 09:09:06.045387 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-852td\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:06 crc kubenswrapper[4944]: I1124 09:09:06.062356 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqhfn\" (UniqueName: \"kubernetes.io/projected/eab6b2a1-815c-473a-ac11-b10194ab0e37-kube-api-access-qqhfn\") pod \"dnsmasq-dns-78dd6ddcc-852td\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:06 crc kubenswrapper[4944]: I1124 09:09:06.137920 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" Nov 24 09:09:06 crc kubenswrapper[4944]: I1124 09:09:06.209069 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:06 crc kubenswrapper[4944]: I1124 09:09:06.573734 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fdkn4"] Nov 24 09:09:06 crc kubenswrapper[4944]: W1124 09:09:06.576634 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7d15acb_9ec0_4cc2_bf76_e379e67b59b1.slice/crio-589a7020f8de42d767bcd96ed9cf59b3072cecbd5e12a627e18e769491e82503 WatchSource:0}: Error finding container 589a7020f8de42d767bcd96ed9cf59b3072cecbd5e12a627e18e769491e82503: Status 404 returned error can't find the container with id 589a7020f8de42d767bcd96ed9cf59b3072cecbd5e12a627e18e769491e82503 Nov 24 09:09:06 crc kubenswrapper[4944]: I1124 09:09:06.666318 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-852td"] Nov 24 09:09:06 crc kubenswrapper[4944]: W1124 09:09:06.669345 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeab6b2a1_815c_473a_ac11_b10194ab0e37.slice/crio-150d8aad8035135f2b178e1f76a76b88219a3261d9a1d95415e50b09538ba661 WatchSource:0}: Error finding container 150d8aad8035135f2b178e1f76a76b88219a3261d9a1d95415e50b09538ba661: Status 404 returned error can't find the container with id 150d8aad8035135f2b178e1f76a76b88219a3261d9a1d95415e50b09538ba661 Nov 24 09:09:06 crc kubenswrapper[4944]: I1124 09:09:06.673887 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" event={"ID":"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1","Type":"ContainerStarted","Data":"589a7020f8de42d767bcd96ed9cf59b3072cecbd5e12a627e18e769491e82503"} Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.618483 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fdkn4"] Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.673076 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6dkvq"] Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.674347 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.680653 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6dkvq"] Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.711066 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-852td" event={"ID":"eab6b2a1-815c-473a-ac11-b10194ab0e37","Type":"ContainerStarted","Data":"150d8aad8035135f2b178e1f76a76b88219a3261d9a1d95415e50b09538ba661"} Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.785078 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpdfj\" (UniqueName: \"kubernetes.io/projected/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-kube-api-access-mpdfj\") pod \"dnsmasq-dns-5ccc8479f9-6dkvq\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.785160 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-6dkvq\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.785342 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-config\") pod \"dnsmasq-dns-5ccc8479f9-6dkvq\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.886848 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-config\") pod \"dnsmasq-dns-5ccc8479f9-6dkvq\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.886976 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpdfj\" (UniqueName: \"kubernetes.io/projected/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-kube-api-access-mpdfj\") pod \"dnsmasq-dns-5ccc8479f9-6dkvq\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.887139 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-6dkvq\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.888856 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-config\") pod \"dnsmasq-dns-5ccc8479f9-6dkvq\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.889336 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-6dkvq\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:07 crc kubenswrapper[4944]: I1124 09:09:07.913979 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpdfj\" (UniqueName: \"kubernetes.io/projected/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-kube-api-access-mpdfj\") pod \"dnsmasq-dns-5ccc8479f9-6dkvq\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.018729 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.235200 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-852td"] Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.265597 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5mj5n"] Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.267376 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.293294 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5mj5n\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.293341 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-config\") pod \"dnsmasq-dns-57d769cc4f-5mj5n\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.293428 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnh4m\" (UniqueName: \"kubernetes.io/projected/67127c77-12a7-41bd-a975-863417818765-kube-api-access-nnh4m\") pod \"dnsmasq-dns-57d769cc4f-5mj5n\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.302796 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5mj5n"] Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.395149 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5mj5n\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.395201 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-config\") pod \"dnsmasq-dns-57d769cc4f-5mj5n\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.395259 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnh4m\" (UniqueName: \"kubernetes.io/projected/67127c77-12a7-41bd-a975-863417818765-kube-api-access-nnh4m\") pod \"dnsmasq-dns-57d769cc4f-5mj5n\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.396777 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5mj5n\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.397016 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-config\") pod \"dnsmasq-dns-57d769cc4f-5mj5n\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.437375 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnh4m\" (UniqueName: \"kubernetes.io/projected/67127c77-12a7-41bd-a975-863417818765-kube-api-access-nnh4m\") pod \"dnsmasq-dns-57d769cc4f-5mj5n\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.595959 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6dkvq"] Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.598345 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.724895 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" event={"ID":"d14b3bac-81cb-4c72-a8bf-6bbfd829575f","Type":"ContainerStarted","Data":"662d52bae705e0648d25f4b727ccb945185d738badecbb64047e24b2fc2dccb8"} Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.803875 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.805233 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.808529 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.808753 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.808880 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.809020 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4d67l" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.809191 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.809247 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.809197 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.842545 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.905669 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.905720 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgfx2\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-kube-api-access-jgfx2\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.905752 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.905804 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.905853 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.905876 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.906000 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.906112 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.906253 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.906349 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:08 crc kubenswrapper[4944]: I1124 09:09:08.906408 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.008563 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.008615 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgfx2\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-kube-api-access-jgfx2\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.008649 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.008683 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.008722 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.008744 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.008777 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.008846 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.008888 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.009324 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.009867 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.009902 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.010216 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.011426 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.013016 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.013288 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.013364 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.014527 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.014740 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.014917 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.015908 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.029272 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgfx2\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-kube-api-access-jgfx2\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.032389 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.149155 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.191066 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5mj5n"] Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.389702 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.391635 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.394556 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.394747 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.394982 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-prx46" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.395689 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.396031 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.396452 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.397853 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.405325 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.432414 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz227\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-kube-api-access-wz227\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.432460 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.432494 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.432522 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.432575 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.432600 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/67e92d3f-3532-467f-bf38-c9c3107d4fcb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.432612 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.432645 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.432662 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.432698 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.432729 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/67e92d3f-3532-467f-bf38-c9c3107d4fcb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.537564 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz227\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-kube-api-access-wz227\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.537613 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.537643 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.537672 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.537711 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.537743 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/67e92d3f-3532-467f-bf38-c9c3107d4fcb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.537765 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.537797 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.537822 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.537861 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.537904 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/67e92d3f-3532-467f-bf38-c9c3107d4fcb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.539345 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.540381 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.540390 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.540703 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.541178 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.545332 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.545748 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/67e92d3f-3532-467f-bf38-c9c3107d4fcb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.546545 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.556728 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz227\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-kube-api-access-wz227\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.562880 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.567762 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.570010 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/67e92d3f-3532-467f-bf38-c9c3107d4fcb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.726107 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.770123 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" event={"ID":"67127c77-12a7-41bd-a975-863417818765","Type":"ContainerStarted","Data":"d73ec33895f8f79b2c825ef4b432eed1043b02ad68be662b8639bc9fee1e5a05"} Nov 24 09:09:09 crc kubenswrapper[4944]: I1124 09:09:09.803424 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.418162 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 09:09:10 crc kubenswrapper[4944]: W1124 09:09:10.424949 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67e92d3f_3532_467f_bf38_c9c3107d4fcb.slice/crio-31533e8484a6664f1f44ac64cf5fa6cf56f3bd282b02782b02cbcfbce6fe1aed WatchSource:0}: Error finding container 31533e8484a6664f1f44ac64cf5fa6cf56f3bd282b02782b02cbcfbce6fe1aed: Status 404 returned error can't find the container with id 31533e8484a6664f1f44ac64cf5fa6cf56f3bd282b02782b02cbcfbce6fe1aed Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.780457 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.794648 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.796893 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.798517 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.805413 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.805664 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-t8kqj" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.810836 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.811915 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"67e92d3f-3532-467f-bf38-c9c3107d4fcb","Type":"ContainerStarted","Data":"31533e8484a6664f1f44ac64cf5fa6cf56f3bd282b02782b02cbcfbce6fe1aed"} Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.813818 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5","Type":"ContainerStarted","Data":"47d0fc0355d2949def167fe23b4bb446e7348b2e8b764656ab97634732468cbc"} Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.828279 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.886762 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.886833 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.886854 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-generated\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.886878 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.886916 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kolla-config\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.886941 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-default\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.886969 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsg45\" (UniqueName: \"kubernetes.io/projected/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kube-api-access-fsg45\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.887002 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-operator-scripts\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.988276 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.988338 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-generated\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.988382 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.988417 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kolla-config\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.988474 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-default\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.988514 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsg45\" (UniqueName: \"kubernetes.io/projected/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kube-api-access-fsg45\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.988568 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-operator-scripts\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.988589 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.988996 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.989665 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-generated\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.990976 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-default\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.991830 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-operator-scripts\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.992433 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kolla-config\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:10 crc kubenswrapper[4944]: I1124 09:09:10.998115 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.012620 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.018552 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsg45\" (UniqueName: \"kubernetes.io/projected/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kube-api-access-fsg45\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.020889 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " pod="openstack/openstack-galera-0" Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.138734 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.733285 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.831961 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"18d26ea6-eff1-4ad5-b50b-dd233a22e982","Type":"ContainerStarted","Data":"5485cef22f5049df9ee59e90c4a019327f2dc0ec0af74eca05327bfcb5d47245"} Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.918179 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.919657 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.927501 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.927732 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.927929 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-fcj6l" Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.929182 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 09:09:11 crc kubenswrapper[4944]: I1124 09:09:11.957176 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.016909 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.017025 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.017108 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.017157 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e454a738-0872-41d8-9432-17432276248c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.017183 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.017211 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.017239 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp7ql\" (UniqueName: \"kubernetes.io/projected/e454a738-0872-41d8-9432-17432276248c-kube-api-access-tp7ql\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.017272 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.118455 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.118545 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e454a738-0872-41d8-9432-17432276248c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.118573 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.119605 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e454a738-0872-41d8-9432-17432276248c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.119688 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.119727 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp7ql\" (UniqueName: \"kubernetes.io/projected/e454a738-0872-41d8-9432-17432276248c-kube-api-access-tp7ql\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.119758 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.119832 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.120373 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.120635 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.122558 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.123272 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.123410 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.140029 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.147180 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp7ql\" (UniqueName: \"kubernetes.io/projected/e454a738-0872-41d8-9432-17432276248c-kube-api-access-tp7ql\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.148236 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.165272 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.292235 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.397178 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.399705 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.399875 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.403457 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-zkwv4" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.403720 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.407409 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.540564 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-config-data\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.542290 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jrgm\" (UniqueName: \"kubernetes.io/projected/35137b92-4f56-4399-8121-89bdb126162b-kube-api-access-8jrgm\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.542343 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.542577 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-kolla-config\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.542601 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.643683 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-config-data\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.643769 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jrgm\" (UniqueName: \"kubernetes.io/projected/35137b92-4f56-4399-8121-89bdb126162b-kube-api-access-8jrgm\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.643809 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.643874 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-kolla-config\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.643894 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.646499 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-kolla-config\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.647068 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-config-data\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.655720 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.677194 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.699254 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jrgm\" (UniqueName: \"kubernetes.io/projected/35137b92-4f56-4399-8121-89bdb126162b-kube-api-access-8jrgm\") pod \"memcached-0\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.814817 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 09:09:12 crc kubenswrapper[4944]: I1124 09:09:12.993839 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 09:09:13 crc kubenswrapper[4944]: I1124 09:09:13.343336 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 09:09:13 crc kubenswrapper[4944]: I1124 09:09:13.994126 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:09:13 crc kubenswrapper[4944]: I1124 09:09:13.995813 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 09:09:14 crc kubenswrapper[4944]: I1124 09:09:14.000032 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-8wbz9" Nov 24 09:09:14 crc kubenswrapper[4944]: I1124 09:09:14.001210 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:09:14 crc kubenswrapper[4944]: I1124 09:09:14.175858 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpxqs\" (UniqueName: \"kubernetes.io/projected/ca22541f-1cf6-4ddf-9da8-23faad6ef99e-kube-api-access-bpxqs\") pod \"kube-state-metrics-0\" (UID: \"ca22541f-1cf6-4ddf-9da8-23faad6ef99e\") " pod="openstack/kube-state-metrics-0" Nov 24 09:09:14 crc kubenswrapper[4944]: I1124 09:09:14.277555 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpxqs\" (UniqueName: \"kubernetes.io/projected/ca22541f-1cf6-4ddf-9da8-23faad6ef99e-kube-api-access-bpxqs\") pod \"kube-state-metrics-0\" (UID: \"ca22541f-1cf6-4ddf-9da8-23faad6ef99e\") " pod="openstack/kube-state-metrics-0" Nov 24 09:09:14 crc kubenswrapper[4944]: I1124 09:09:14.314324 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpxqs\" (UniqueName: \"kubernetes.io/projected/ca22541f-1cf6-4ddf-9da8-23faad6ef99e-kube-api-access-bpxqs\") pod \"kube-state-metrics-0\" (UID: \"ca22541f-1cf6-4ddf-9da8-23faad6ef99e\") " pod="openstack/kube-state-metrics-0" Nov 24 09:09:14 crc kubenswrapper[4944]: I1124 09:09:14.333758 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 09:09:16 crc kubenswrapper[4944]: I1124 09:09:16.966875 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7gtjj"] Nov 24 09:09:16 crc kubenswrapper[4944]: I1124 09:09:16.969000 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:16 crc kubenswrapper[4944]: I1124 09:09:16.969880 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-npmfc"] Nov 24 09:09:16 crc kubenswrapper[4944]: I1124 09:09:16.970881 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc" Nov 24 09:09:16 crc kubenswrapper[4944]: I1124 09:09:16.974024 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-cwtxx" Nov 24 09:09:16 crc kubenswrapper[4944]: I1124 09:09:16.985829 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 09:09:16 crc kubenswrapper[4944]: I1124 09:09:16.986576 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 09:09:16 crc kubenswrapper[4944]: I1124 09:09:16.996917 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-npmfc"] Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.017980 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7gtjj"] Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122135 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-lib\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122179 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-log\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122205 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrvn4\" (UniqueName: \"kubernetes.io/projected/69084f7b-b347-4fdf-917e-2f534f3cc47c-kube-api-access-qrvn4\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122225 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run-ovn\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122341 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-run\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122396 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69084f7b-b347-4fdf-917e-2f534f3cc47c-scripts\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122443 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-etc-ovs\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122473 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-scripts\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122525 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-ovn-controller-tls-certs\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122592 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122608 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-combined-ca-bundle\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122627 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-log-ovn\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.122661 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9htrv\" (UniqueName: \"kubernetes.io/projected/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-kube-api-access-9htrv\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.225745 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-ovn-controller-tls-certs\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.225953 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.225989 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-log-ovn\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226019 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-combined-ca-bundle\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226068 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9htrv\" (UniqueName: \"kubernetes.io/projected/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-kube-api-access-9htrv\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226108 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-lib\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226155 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-log\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226183 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrvn4\" (UniqueName: \"kubernetes.io/projected/69084f7b-b347-4fdf-917e-2f534f3cc47c-kube-api-access-qrvn4\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226220 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run-ovn\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226257 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-run\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226284 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69084f7b-b347-4fdf-917e-2f534f3cc47c-scripts\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226317 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-etc-ovs\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226347 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-scripts\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226794 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226891 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-run\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226933 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-log\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.226975 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run-ovn\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.227059 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-lib\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.227127 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-log-ovn\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.227118 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-etc-ovs\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.229655 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69084f7b-b347-4fdf-917e-2f534f3cc47c-scripts\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.229952 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-scripts\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.236798 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-combined-ca-bundle\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.236932 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-ovn-controller-tls-certs\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.249563 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9htrv\" (UniqueName: \"kubernetes.io/projected/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-kube-api-access-9htrv\") pod \"ovn-controller-ovs-7gtjj\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.253866 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrvn4\" (UniqueName: \"kubernetes.io/projected/69084f7b-b347-4fdf-917e-2f534f3cc47c-kube-api-access-qrvn4\") pod \"ovn-controller-npmfc\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " pod="openstack/ovn-controller-npmfc" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.292574 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:17 crc kubenswrapper[4944]: I1124 09:09:17.300888 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.431122 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.435565 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.440263 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.440850 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.441072 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.441266 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-7pkqm" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.441387 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.442176 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.573243 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.573334 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.573376 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.573399 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-config\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.573424 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.573728 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.573823 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.573867 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h4qw\" (UniqueName: \"kubernetes.io/projected/b9d45d4c-ab9f-426c-a193-7f46c398ed64-kube-api-access-4h4qw\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.675908 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.675998 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.676034 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.676083 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h4qw\" (UniqueName: \"kubernetes.io/projected/b9d45d4c-ab9f-426c-a193-7f46c398ed64-kube-api-access-4h4qw\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.676112 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.676166 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.676206 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.676225 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-config\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.676647 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.676685 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.677452 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-config\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.677789 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.687600 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.697936 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h4qw\" (UniqueName: \"kubernetes.io/projected/b9d45d4c-ab9f-426c-a193-7f46c398ed64-kube-api-access-4h4qw\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.709585 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.715313 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.723964 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.788804 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:18 crc kubenswrapper[4944]: I1124 09:09:18.942705 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e454a738-0872-41d8-9432-17432276248c","Type":"ContainerStarted","Data":"6236aed52bc908141599bf898d07bc26522c296edb4567ff1515a8486312f72f"} Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.560555 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.563723 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.565938 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vjsdm" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.566422 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.566500 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.566601 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.572118 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.731780 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.731866 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.732191 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4xgf\" (UniqueName: \"kubernetes.io/projected/2c04aca3-abaa-4c62-8e62-af920276cc50-kube-api-access-r4xgf\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.732279 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.732384 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.732493 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.732543 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-config\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.732579 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.834571 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.834667 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4xgf\" (UniqueName: \"kubernetes.io/projected/2c04aca3-abaa-4c62-8e62-af920276cc50-kube-api-access-r4xgf\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.834695 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.834727 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.834754 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.834772 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-config\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.834791 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.834810 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.835141 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.836554 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-config\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.836942 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.837867 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.847513 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.848230 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.852372 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.852798 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4xgf\" (UniqueName: \"kubernetes.io/projected/2c04aca3-abaa-4c62-8e62-af920276cc50-kube-api-access-r4xgf\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.863700 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:21 crc kubenswrapper[4944]: I1124 09:09:21.900230 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:22 crc kubenswrapper[4944]: I1124 09:09:22.979851 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"35137b92-4f56-4399-8121-89bdb126162b","Type":"ContainerStarted","Data":"646c15a367bff50291e941fd08d418827b83f1e8278127d694dff9bd10c89f4a"} Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.152097 4944 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.152787 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6cncj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-fdkn4_openstack(b7d15acb-9ec0-4cc2-bf76-e379e67b59b1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.154287 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" podUID="b7d15acb-9ec0-4cc2-bf76-e379e67b59b1" Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.203384 4944 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.203561 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mpdfj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-6dkvq_openstack(d14b3bac-81cb-4c72-a8bf-6bbfd829575f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.205163 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" podUID="d14b3bac-81cb-4c72-a8bf-6bbfd829575f" Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.246010 4944 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.246417 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nnh4m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-5mj5n_openstack(67127c77-12a7-41bd-a975-863417818765): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.247628 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" podUID="67127c77-12a7-41bd-a975-863417818765" Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.273287 4944 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.273437 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qqhfn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-852td_openstack(eab6b2a1-815c-473a-ac11-b10194ab0e37): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 09:09:28 crc kubenswrapper[4944]: E1124 09:09:28.274663 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-852td" podUID="eab6b2a1-815c-473a-ac11-b10194ab0e37" Nov 24 09:09:28 crc kubenswrapper[4944]: I1124 09:09:28.543646 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-npmfc"] Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.063605 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-npmfc" event={"ID":"69084f7b-b347-4fdf-917e-2f534f3cc47c","Type":"ContainerStarted","Data":"9f239ec933ea933cd803c9ec051a6a8fa85566e266369c4d911776e1e04db753"} Nov 24 09:09:29 crc kubenswrapper[4944]: E1124 09:09:29.066308 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" podUID="67127c77-12a7-41bd-a975-863417818765" Nov 24 09:09:29 crc kubenswrapper[4944]: E1124 09:09:29.067075 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" podUID="d14b3bac-81cb-4c72-a8bf-6bbfd829575f" Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.461727 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7gtjj"] Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.654569 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.671226 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.710915 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.757910 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-dns-svc\") pod \"eab6b2a1-815c-473a-ac11-b10194ab0e37\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.758990 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-config\") pod \"eab6b2a1-815c-473a-ac11-b10194ab0e37\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.759056 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-config\") pod \"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1\" (UID: \"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1\") " Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.759098 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cncj\" (UniqueName: \"kubernetes.io/projected/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-kube-api-access-6cncj\") pod \"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1\" (UID: \"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1\") " Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.759140 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqhfn\" (UniqueName: \"kubernetes.io/projected/eab6b2a1-815c-473a-ac11-b10194ab0e37-kube-api-access-qqhfn\") pod \"eab6b2a1-815c-473a-ac11-b10194ab0e37\" (UID: \"eab6b2a1-815c-473a-ac11-b10194ab0e37\") " Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.758518 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eab6b2a1-815c-473a-ac11-b10194ab0e37" (UID: "eab6b2a1-815c-473a-ac11-b10194ab0e37"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.759435 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-config" (OuterVolumeSpecName: "config") pod "eab6b2a1-815c-473a-ac11-b10194ab0e37" (UID: "eab6b2a1-815c-473a-ac11-b10194ab0e37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.759676 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-config" (OuterVolumeSpecName: "config") pod "b7d15acb-9ec0-4cc2-bf76-e379e67b59b1" (UID: "b7d15acb-9ec0-4cc2-bf76-e379e67b59b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.761705 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.761724 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.761733 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eab6b2a1-815c-473a-ac11-b10194ab0e37-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.794581 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.797524 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eab6b2a1-815c-473a-ac11-b10194ab0e37-kube-api-access-qqhfn" (OuterVolumeSpecName: "kube-api-access-qqhfn") pod "eab6b2a1-815c-473a-ac11-b10194ab0e37" (UID: "eab6b2a1-815c-473a-ac11-b10194ab0e37"). InnerVolumeSpecName "kube-api-access-qqhfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.797971 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-kube-api-access-6cncj" (OuterVolumeSpecName: "kube-api-access-6cncj") pod "b7d15acb-9ec0-4cc2-bf76-e379e67b59b1" (UID: "b7d15acb-9ec0-4cc2-bf76-e379e67b59b1"). InnerVolumeSpecName "kube-api-access-6cncj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:29 crc kubenswrapper[4944]: W1124 09:09:29.800240 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca22541f_1cf6_4ddf_9da8_23faad6ef99e.slice/crio-6eaf15f7622eac875b82f0db2a8c6ddb4dee3321c2a7ba16ab3010bb51e08e02 WatchSource:0}: Error finding container 6eaf15f7622eac875b82f0db2a8c6ddb4dee3321c2a7ba16ab3010bb51e08e02: Status 404 returned error can't find the container with id 6eaf15f7622eac875b82f0db2a8c6ddb4dee3321c2a7ba16ab3010bb51e08e02 Nov 24 09:09:29 crc kubenswrapper[4944]: W1124 09:09:29.800766 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c04aca3_abaa_4c62_8e62_af920276cc50.slice/crio-3cc7f342b1dff7530f9912e2834f34fe4b434b6e60ff2bc0acd8adfe099bd5bf WatchSource:0}: Error finding container 3cc7f342b1dff7530f9912e2834f34fe4b434b6e60ff2bc0acd8adfe099bd5bf: Status 404 returned error can't find the container with id 3cc7f342b1dff7530f9912e2834f34fe4b434b6e60ff2bc0acd8adfe099bd5bf Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.863754 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cncj\" (UniqueName: \"kubernetes.io/projected/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1-kube-api-access-6cncj\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:29 crc kubenswrapper[4944]: I1124 09:09:29.863799 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqhfn\" (UniqueName: \"kubernetes.io/projected/eab6b2a1-815c-473a-ac11-b10194ab0e37-kube-api-access-qqhfn\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.039119 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.074347 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"18d26ea6-eff1-4ad5-b50b-dd233a22e982","Type":"ContainerStarted","Data":"f55d411db821167a7793432c6ee6466117e2082720794f9b45e02915d1b9298e"} Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.080432 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7gtjj" event={"ID":"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b","Type":"ContainerStarted","Data":"6ec02dfe04c23880d4b3c99aa2deaa2b332f08aad186936b4a51d23a76d9db5d"} Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.082426 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2c04aca3-abaa-4c62-8e62-af920276cc50","Type":"ContainerStarted","Data":"3cc7f342b1dff7530f9912e2834f34fe4b434b6e60ff2bc0acd8adfe099bd5bf"} Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.085843 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e454a738-0872-41d8-9432-17432276248c","Type":"ContainerStarted","Data":"f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c"} Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.089926 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-852td" event={"ID":"eab6b2a1-815c-473a-ac11-b10194ab0e37","Type":"ContainerDied","Data":"150d8aad8035135f2b178e1f76a76b88219a3261d9a1d95415e50b09538ba661"} Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.090028 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-852td" Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.093789 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"35137b92-4f56-4399-8121-89bdb126162b","Type":"ContainerStarted","Data":"10473e2e58eff197e7d6054c78b1b43c310d54ae418a52fb6168c9c0b59133e8"} Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.094839 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.097031 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ca22541f-1cf6-4ddf-9da8-23faad6ef99e","Type":"ContainerStarted","Data":"6eaf15f7622eac875b82f0db2a8c6ddb4dee3321c2a7ba16ab3010bb51e08e02"} Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.100956 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" event={"ID":"b7d15acb-9ec0-4cc2-bf76-e379e67b59b1","Type":"ContainerDied","Data":"589a7020f8de42d767bcd96ed9cf59b3072cecbd5e12a627e18e769491e82503"} Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.101084 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fdkn4" Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.162897 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=11.778948433 podStartE2EDuration="18.162876399s" podCreationTimestamp="2025-11-24 09:09:12 +0000 UTC" firstStartedPulling="2025-11-24 09:09:22.539579913 +0000 UTC m=+1023.074020375" lastFinishedPulling="2025-11-24 09:09:28.923507879 +0000 UTC m=+1029.457948341" observedRunningTime="2025-11-24 09:09:30.152433326 +0000 UTC m=+1030.686873808" watchObservedRunningTime="2025-11-24 09:09:30.162876399 +0000 UTC m=+1030.697316861" Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.207207 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fdkn4"] Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.215889 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fdkn4"] Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.252600 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-852td"] Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.266200 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-852td"] Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.286864 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7d15acb-9ec0-4cc2-bf76-e379e67b59b1" path="/var/lib/kubelet/pods/b7d15acb-9ec0-4cc2-bf76-e379e67b59b1/volumes" Nov 24 09:09:30 crc kubenswrapper[4944]: I1124 09:09:30.287522 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eab6b2a1-815c-473a-ac11-b10194ab0e37" path="/var/lib/kubelet/pods/eab6b2a1-815c-473a-ac11-b10194ab0e37/volumes" Nov 24 09:09:31 crc kubenswrapper[4944]: I1124 09:09:31.110964 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b9d45d4c-ab9f-426c-a193-7f46c398ed64","Type":"ContainerStarted","Data":"c38d9bc8e2ba464c90202d5ba1fe348490abf6a227c1249ebfab53d71639a12d"} Nov 24 09:09:34 crc kubenswrapper[4944]: I1124 09:09:34.134140 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"67e92d3f-3532-467f-bf38-c9c3107d4fcb","Type":"ContainerStarted","Data":"49cd4f1b400eddca6bdd6df980eb6336eb7f945c92b13a7f0c34e97beb31f078"} Nov 24 09:09:34 crc kubenswrapper[4944]: I1124 09:09:34.135953 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5","Type":"ContainerStarted","Data":"0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033"} Nov 24 09:09:36 crc kubenswrapper[4944]: I1124 09:09:36.152152 4944 generic.go:334] "Generic (PLEG): container finished" podID="e454a738-0872-41d8-9432-17432276248c" containerID="f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c" exitCode=0 Nov 24 09:09:36 crc kubenswrapper[4944]: I1124 09:09:36.152270 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e454a738-0872-41d8-9432-17432276248c","Type":"ContainerDied","Data":"f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c"} Nov 24 09:09:36 crc kubenswrapper[4944]: I1124 09:09:36.155714 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-npmfc" event={"ID":"69084f7b-b347-4fdf-917e-2f534f3cc47c","Type":"ContainerStarted","Data":"b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009"} Nov 24 09:09:36 crc kubenswrapper[4944]: I1124 09:09:36.156239 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-npmfc" Nov 24 09:09:36 crc kubenswrapper[4944]: I1124 09:09:36.158798 4944 generic.go:334] "Generic (PLEG): container finished" podID="18d26ea6-eff1-4ad5-b50b-dd233a22e982" containerID="f55d411db821167a7793432c6ee6466117e2082720794f9b45e02915d1b9298e" exitCode=0 Nov 24 09:09:36 crc kubenswrapper[4944]: I1124 09:09:36.158854 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"18d26ea6-eff1-4ad5-b50b-dd233a22e982","Type":"ContainerDied","Data":"f55d411db821167a7793432c6ee6466117e2082720794f9b45e02915d1b9298e"} Nov 24 09:09:36 crc kubenswrapper[4944]: I1124 09:09:36.160937 4944 generic.go:334] "Generic (PLEG): container finished" podID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerID="6a3410c6257d72bcea62d39ce6013b2c346a2c2f1e11397a1e28d3f46511dc6e" exitCode=0 Nov 24 09:09:36 crc kubenswrapper[4944]: I1124 09:09:36.160969 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7gtjj" event={"ID":"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b","Type":"ContainerDied","Data":"6a3410c6257d72bcea62d39ce6013b2c346a2c2f1e11397a1e28d3f46511dc6e"} Nov 24 09:09:36 crc kubenswrapper[4944]: I1124 09:09:36.214687 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-npmfc" podStartSLOduration=13.85323197 podStartE2EDuration="20.214670759s" podCreationTimestamp="2025-11-24 09:09:16 +0000 UTC" firstStartedPulling="2025-11-24 09:09:28.891632822 +0000 UTC m=+1029.426073284" lastFinishedPulling="2025-11-24 09:09:35.253071621 +0000 UTC m=+1035.787512073" observedRunningTime="2025-11-24 09:09:36.21000075 +0000 UTC m=+1036.744441232" watchObservedRunningTime="2025-11-24 09:09:36.214670759 +0000 UTC m=+1036.749111221" Nov 24 09:09:37 crc kubenswrapper[4944]: I1124 09:09:37.168320 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7gtjj" event={"ID":"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b","Type":"ContainerStarted","Data":"35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb"} Nov 24 09:09:37 crc kubenswrapper[4944]: I1124 09:09:37.170093 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2c04aca3-abaa-4c62-8e62-af920276cc50","Type":"ContainerStarted","Data":"61ee3036c3b98fca180967e52f25b5128d65397ebffba1911bb8a5e25c2ec958"} Nov 24 09:09:37 crc kubenswrapper[4944]: I1124 09:09:37.172240 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e454a738-0872-41d8-9432-17432276248c","Type":"ContainerStarted","Data":"06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d"} Nov 24 09:09:37 crc kubenswrapper[4944]: I1124 09:09:37.173801 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ca22541f-1cf6-4ddf-9da8-23faad6ef99e","Type":"ContainerStarted","Data":"27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e"} Nov 24 09:09:37 crc kubenswrapper[4944]: I1124 09:09:37.173935 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 09:09:37 crc kubenswrapper[4944]: I1124 09:09:37.176849 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"18d26ea6-eff1-4ad5-b50b-dd233a22e982","Type":"ContainerStarted","Data":"1a95527c06d2048bd5100167ef2d1b762657135708fcae5c22fd4931034b3057"} Nov 24 09:09:37 crc kubenswrapper[4944]: I1124 09:09:37.178716 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b9d45d4c-ab9f-426c-a193-7f46c398ed64","Type":"ContainerStarted","Data":"5ef21e16c9d939c2affb5458115459c6969a680e9f313ed1d7b13c1de4f3f96a"} Nov 24 09:09:37 crc kubenswrapper[4944]: I1124 09:09:37.197633 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=16.371168372 podStartE2EDuration="27.197613947s" podCreationTimestamp="2025-11-24 09:09:10 +0000 UTC" firstStartedPulling="2025-11-24 09:09:18.070986472 +0000 UTC m=+1018.605426934" lastFinishedPulling="2025-11-24 09:09:28.897432047 +0000 UTC m=+1029.431872509" observedRunningTime="2025-11-24 09:09:37.19236947 +0000 UTC m=+1037.726809952" watchObservedRunningTime="2025-11-24 09:09:37.197613947 +0000 UTC m=+1037.732054409" Nov 24 09:09:37 crc kubenswrapper[4944]: I1124 09:09:37.212926 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=11.732254372 podStartE2EDuration="28.212911075s" podCreationTimestamp="2025-11-24 09:09:09 +0000 UTC" firstStartedPulling="2025-11-24 09:09:11.740060067 +0000 UTC m=+1012.274500529" lastFinishedPulling="2025-11-24 09:09:28.22071677 +0000 UTC m=+1028.755157232" observedRunningTime="2025-11-24 09:09:37.211571242 +0000 UTC m=+1037.746011714" watchObservedRunningTime="2025-11-24 09:09:37.212911075 +0000 UTC m=+1037.747351537" Nov 24 09:09:37 crc kubenswrapper[4944]: I1124 09:09:37.232690 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=17.149986569 podStartE2EDuration="24.232668975s" podCreationTimestamp="2025-11-24 09:09:13 +0000 UTC" firstStartedPulling="2025-11-24 09:09:29.803169695 +0000 UTC m=+1030.337610157" lastFinishedPulling="2025-11-24 09:09:36.885852101 +0000 UTC m=+1037.420292563" observedRunningTime="2025-11-24 09:09:37.225079152 +0000 UTC m=+1037.759519614" watchObservedRunningTime="2025-11-24 09:09:37.232668975 +0000 UTC m=+1037.767109437" Nov 24 09:09:37 crc kubenswrapper[4944]: I1124 09:09:37.816270 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 09:09:38 crc kubenswrapper[4944]: I1124 09:09:38.190308 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7gtjj" event={"ID":"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b","Type":"ContainerStarted","Data":"c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac"} Nov 24 09:09:38 crc kubenswrapper[4944]: I1124 09:09:38.190989 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:38 crc kubenswrapper[4944]: I1124 09:09:38.214270 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7gtjj" podStartSLOduration=16.429112297 podStartE2EDuration="22.214255499s" podCreationTimestamp="2025-11-24 09:09:16 +0000 UTC" firstStartedPulling="2025-11-24 09:09:29.467615869 +0000 UTC m=+1030.002056331" lastFinishedPulling="2025-11-24 09:09:35.252759071 +0000 UTC m=+1035.787199533" observedRunningTime="2025-11-24 09:09:38.212542015 +0000 UTC m=+1038.746982477" watchObservedRunningTime="2025-11-24 09:09:38.214255499 +0000 UTC m=+1038.748695961" Nov 24 09:09:39 crc kubenswrapper[4944]: I1124 09:09:39.199895 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:09:41 crc kubenswrapper[4944]: E1124 09:09:41.135127 4944 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.15:40712->38.129.56.15:34895: read tcp 38.129.56.15:40712->38.129.56.15:34895: read: connection reset by peer Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.139379 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.139495 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.217203 4944 generic.go:334] "Generic (PLEG): container finished" podID="67127c77-12a7-41bd-a975-863417818765" containerID="434592ad9cf533bd4d47b291b28cce6adadd5b13a285800b305de9f00b8ecbaf" exitCode=0 Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.217289 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" event={"ID":"67127c77-12a7-41bd-a975-863417818765","Type":"ContainerDied","Data":"434592ad9cf533bd4d47b291b28cce6adadd5b13a285800b305de9f00b8ecbaf"} Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.220858 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b9d45d4c-ab9f-426c-a193-7f46c398ed64","Type":"ContainerStarted","Data":"3ce79850e6073ff3d7e81da5fd81b8be25dfceb1d21948518006ccbe0b1eda7a"} Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.223075 4944 generic.go:334] "Generic (PLEG): container finished" podID="d14b3bac-81cb-4c72-a8bf-6bbfd829575f" containerID="24979742899bd5b53f8ded8da1e731c34e36522cfe9eb536f8b735d53e121eb6" exitCode=0 Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.223167 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" event={"ID":"d14b3bac-81cb-4c72-a8bf-6bbfd829575f","Type":"ContainerDied","Data":"24979742899bd5b53f8ded8da1e731c34e36522cfe9eb536f8b735d53e121eb6"} Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.226104 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2c04aca3-abaa-4c62-8e62-af920276cc50","Type":"ContainerStarted","Data":"878cbf2bbb403dd3a431aa6956e733ae81fc764310e89c3f03f0d40d6516dc09"} Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.280210 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=10.711647489 podStartE2EDuration="21.280186995s" podCreationTimestamp="2025-11-24 09:09:20 +0000 UTC" firstStartedPulling="2025-11-24 09:09:29.80461129 +0000 UTC m=+1030.339051752" lastFinishedPulling="2025-11-24 09:09:40.373150796 +0000 UTC m=+1040.907591258" observedRunningTime="2025-11-24 09:09:41.274111331 +0000 UTC m=+1041.808551853" watchObservedRunningTime="2025-11-24 09:09:41.280186995 +0000 UTC m=+1041.814627457" Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.305236 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=14.013487029 podStartE2EDuration="24.305210042s" podCreationTimestamp="2025-11-24 09:09:17 +0000 UTC" firstStartedPulling="2025-11-24 09:09:30.112734491 +0000 UTC m=+1030.647174953" lastFinishedPulling="2025-11-24 09:09:40.404457504 +0000 UTC m=+1040.938897966" observedRunningTime="2025-11-24 09:09:41.294725598 +0000 UTC m=+1041.829166070" watchObservedRunningTime="2025-11-24 09:09:41.305210042 +0000 UTC m=+1041.839650524" Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.848464 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.900914 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:41 crc kubenswrapper[4944]: I1124 09:09:41.912682 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.157187 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-1ab7-account-create-c4ljx"] Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.158207 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1ab7-account-create-c4ljx" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.160501 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.173877 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-1ab7-account-create-c4ljx"] Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.217664 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-wgxhq"] Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.218789 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wgxhq" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.224962 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wgxhq"] Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.235456 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" event={"ID":"d14b3bac-81cb-4c72-a8bf-6bbfd829575f","Type":"ContainerStarted","Data":"64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c"} Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.235713 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.238104 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" event={"ID":"67127c77-12a7-41bd-a975-863417818765","Type":"ContainerStarted","Data":"deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697"} Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.255500 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.255816 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.255972 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" podStartSLOduration=3.488828218 podStartE2EDuration="35.255959444s" podCreationTimestamp="2025-11-24 09:09:07 +0000 UTC" firstStartedPulling="2025-11-24 09:09:08.639875369 +0000 UTC m=+1009.174315831" lastFinishedPulling="2025-11-24 09:09:40.407006595 +0000 UTC m=+1040.941447057" observedRunningTime="2025-11-24 09:09:42.251682548 +0000 UTC m=+1042.786123010" watchObservedRunningTime="2025-11-24 09:09:42.255959444 +0000 UTC m=+1042.790399906" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.272608 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" podStartSLOduration=2.479387886 podStartE2EDuration="34.272590994s" podCreationTimestamp="2025-11-24 09:09:08 +0000 UTC" firstStartedPulling="2025-11-24 09:09:09.219397859 +0000 UTC m=+1009.753838321" lastFinishedPulling="2025-11-24 09:09:41.012600967 +0000 UTC m=+1041.547041429" observedRunningTime="2025-11-24 09:09:42.269022771 +0000 UTC m=+1042.803463243" watchObservedRunningTime="2025-11-24 09:09:42.272590994 +0000 UTC m=+1042.807031456" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.305917 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkrqv\" (UniqueName: \"kubernetes.io/projected/1cc3188b-7842-4aa0-accd-cfa2e303f001-kube-api-access-bkrqv\") pod \"keystone-1ab7-account-create-c4ljx\" (UID: \"1cc3188b-7842-4aa0-accd-cfa2e303f001\") " pod="openstack/keystone-1ab7-account-create-c4ljx" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.306009 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qwkn\" (UniqueName: \"kubernetes.io/projected/d07bba62-af8e-415e-a68d-ed9bf7602bd6-kube-api-access-6qwkn\") pod \"keystone-db-create-wgxhq\" (UID: \"d07bba62-af8e-415e-a68d-ed9bf7602bd6\") " pod="openstack/keystone-db-create-wgxhq" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.306038 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cc3188b-7842-4aa0-accd-cfa2e303f001-operator-scripts\") pod \"keystone-1ab7-account-create-c4ljx\" (UID: \"1cc3188b-7842-4aa0-accd-cfa2e303f001\") " pod="openstack/keystone-1ab7-account-create-c4ljx" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.306141 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d07bba62-af8e-415e-a68d-ed9bf7602bd6-operator-scripts\") pod \"keystone-db-create-wgxhq\" (UID: \"d07bba62-af8e-415e-a68d-ed9bf7602bd6\") " pod="openstack/keystone-db-create-wgxhq" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.319716 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.408345 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkrqv\" (UniqueName: \"kubernetes.io/projected/1cc3188b-7842-4aa0-accd-cfa2e303f001-kube-api-access-bkrqv\") pod \"keystone-1ab7-account-create-c4ljx\" (UID: \"1cc3188b-7842-4aa0-accd-cfa2e303f001\") " pod="openstack/keystone-1ab7-account-create-c4ljx" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.408524 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qwkn\" (UniqueName: \"kubernetes.io/projected/d07bba62-af8e-415e-a68d-ed9bf7602bd6-kube-api-access-6qwkn\") pod \"keystone-db-create-wgxhq\" (UID: \"d07bba62-af8e-415e-a68d-ed9bf7602bd6\") " pod="openstack/keystone-db-create-wgxhq" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.408558 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cc3188b-7842-4aa0-accd-cfa2e303f001-operator-scripts\") pod \"keystone-1ab7-account-create-c4ljx\" (UID: \"1cc3188b-7842-4aa0-accd-cfa2e303f001\") " pod="openstack/keystone-1ab7-account-create-c4ljx" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.408948 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d07bba62-af8e-415e-a68d-ed9bf7602bd6-operator-scripts\") pod \"keystone-db-create-wgxhq\" (UID: \"d07bba62-af8e-415e-a68d-ed9bf7602bd6\") " pod="openstack/keystone-db-create-wgxhq" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.409340 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cc3188b-7842-4aa0-accd-cfa2e303f001-operator-scripts\") pod \"keystone-1ab7-account-create-c4ljx\" (UID: \"1cc3188b-7842-4aa0-accd-cfa2e303f001\") " pod="openstack/keystone-1ab7-account-create-c4ljx" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.410085 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d07bba62-af8e-415e-a68d-ed9bf7602bd6-operator-scripts\") pod \"keystone-db-create-wgxhq\" (UID: \"d07bba62-af8e-415e-a68d-ed9bf7602bd6\") " pod="openstack/keystone-db-create-wgxhq" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.428872 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkrqv\" (UniqueName: \"kubernetes.io/projected/1cc3188b-7842-4aa0-accd-cfa2e303f001-kube-api-access-bkrqv\") pod \"keystone-1ab7-account-create-c4ljx\" (UID: \"1cc3188b-7842-4aa0-accd-cfa2e303f001\") " pod="openstack/keystone-1ab7-account-create-c4ljx" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.429357 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qwkn\" (UniqueName: \"kubernetes.io/projected/d07bba62-af8e-415e-a68d-ed9bf7602bd6-kube-api-access-6qwkn\") pod \"keystone-db-create-wgxhq\" (UID: \"d07bba62-af8e-415e-a68d-ed9bf7602bd6\") " pod="openstack/keystone-db-create-wgxhq" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.458415 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-d4j84"] Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.460029 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4j84" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.469152 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-d4j84"] Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.474969 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1ab7-account-create-c4ljx" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.531963 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-0cdc-account-create-p4c9q"] Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.533633 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0cdc-account-create-p4c9q" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.537229 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wgxhq" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.538218 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.539640 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0cdc-account-create-p4c9q"] Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.613227 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-operator-scripts\") pod \"placement-db-create-d4j84\" (UID: \"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4\") " pod="openstack/placement-db-create-d4j84" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.613594 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zf9p\" (UniqueName: \"kubernetes.io/projected/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-kube-api-access-5zf9p\") pod \"placement-0cdc-account-create-p4c9q\" (UID: \"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80\") " pod="openstack/placement-0cdc-account-create-p4c9q" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.613797 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-operator-scripts\") pod \"placement-0cdc-account-create-p4c9q\" (UID: \"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80\") " pod="openstack/placement-0cdc-account-create-p4c9q" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.613927 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mstp4\" (UniqueName: \"kubernetes.io/projected/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-kube-api-access-mstp4\") pod \"placement-db-create-d4j84\" (UID: \"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4\") " pod="openstack/placement-db-create-d4j84" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.715590 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-operator-scripts\") pod \"placement-db-create-d4j84\" (UID: \"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4\") " pod="openstack/placement-db-create-d4j84" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.715979 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zf9p\" (UniqueName: \"kubernetes.io/projected/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-kube-api-access-5zf9p\") pod \"placement-0cdc-account-create-p4c9q\" (UID: \"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80\") " pod="openstack/placement-0cdc-account-create-p4c9q" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.716014 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-operator-scripts\") pod \"placement-0cdc-account-create-p4c9q\" (UID: \"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80\") " pod="openstack/placement-0cdc-account-create-p4c9q" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.716076 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mstp4\" (UniqueName: \"kubernetes.io/projected/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-kube-api-access-mstp4\") pod \"placement-db-create-d4j84\" (UID: \"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4\") " pod="openstack/placement-db-create-d4j84" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.717163 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-operator-scripts\") pod \"placement-0cdc-account-create-p4c9q\" (UID: \"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80\") " pod="openstack/placement-0cdc-account-create-p4c9q" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.723069 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-operator-scripts\") pod \"placement-db-create-d4j84\" (UID: \"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4\") " pod="openstack/placement-db-create-d4j84" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.733237 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mstp4\" (UniqueName: \"kubernetes.io/projected/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-kube-api-access-mstp4\") pod \"placement-db-create-d4j84\" (UID: \"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4\") " pod="openstack/placement-db-create-d4j84" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.735758 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zf9p\" (UniqueName: \"kubernetes.io/projected/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-kube-api-access-5zf9p\") pod \"placement-0cdc-account-create-p4c9q\" (UID: \"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80\") " pod="openstack/placement-0cdc-account-create-p4c9q" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.789410 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.834713 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.900509 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.939421 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.948080 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4j84" Nov 24 09:09:42 crc kubenswrapper[4944]: I1124 09:09:42.974556 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0cdc-account-create-p4c9q" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.053551 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-1ab7-account-create-c4ljx"] Nov 24 09:09:43 crc kubenswrapper[4944]: W1124 09:09:43.067717 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cc3188b_7842_4aa0_accd_cfa2e303f001.slice/crio-d858c06520b2d55277f3f86d5d9767488799c4dbcb8734327daa45899c4c3b32 WatchSource:0}: Error finding container d858c06520b2d55277f3f86d5d9767488799c4dbcb8734327daa45899c4c3b32: Status 404 returned error can't find the container with id d858c06520b2d55277f3f86d5d9767488799c4dbcb8734327daa45899c4c3b32 Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.130631 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wgxhq"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.258950 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1ab7-account-create-c4ljx" event={"ID":"1cc3188b-7842-4aa0-accd-cfa2e303f001","Type":"ContainerStarted","Data":"d858c06520b2d55277f3f86d5d9767488799c4dbcb8734327daa45899c4c3b32"} Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.266953 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wgxhq" event={"ID":"d07bba62-af8e-415e-a68d-ed9bf7602bd6","Type":"ContainerStarted","Data":"648ecb0fbdc7afb95f8e49dc906cd93b5361b8e0e256bcf322a72c08667e44e3"} Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.268081 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.317904 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.320860 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.365243 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.412178 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-d4j84"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.508865 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-0cdc-account-create-p4c9q"] Nov 24 09:09:43 crc kubenswrapper[4944]: W1124 09:09:43.528810 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9e4a45c_7aa7_4c04_8c86_a7927a7ada80.slice/crio-08d9bc032c492129324fe0b78694ba4d3ea10cf243055e137988cf9e86fd6d5f WatchSource:0}: Error finding container 08d9bc032c492129324fe0b78694ba4d3ea10cf243055e137988cf9e86fd6d5f: Status 404 returned error can't find the container with id 08d9bc032c492129324fe0b78694ba4d3ea10cf243055e137988cf9e86fd6d5f Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.551363 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6dkvq"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.594538 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-srfqr"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.600210 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.604358 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.606692 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-srfqr"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.616824 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.630590 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-config\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.630702 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.630746 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.630813 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmghx\" (UniqueName: \"kubernetes.io/projected/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-kube-api-access-pmghx\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.637256 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-m2b6x"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.638907 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.643744 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.675799 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-m2b6x"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.732332 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovn-rundir\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.732388 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-config\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.732432 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c521e9a3-0d64-4658-8a00-690d20e619d0-config\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.732466 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcfxw\" (UniqueName: \"kubernetes.io/projected/c521e9a3-0d64-4658-8a00-690d20e619d0-kube-api-access-hcfxw\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.732494 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.732530 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.732564 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.732589 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovs-rundir\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.732631 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-combined-ca-bundle\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.732660 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmghx\" (UniqueName: \"kubernetes.io/projected/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-kube-api-access-pmghx\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.733851 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-config\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.734971 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.739399 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.762902 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5mj5n"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.768980 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmghx\" (UniqueName: \"kubernetes.io/projected/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-kube-api-access-pmghx\") pod \"dnsmasq-dns-5bf47b49b7-srfqr\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.780594 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-xsfw4"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.782154 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.788024 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.808443 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xsfw4"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.819603 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.821122 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.822697 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.822809 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-txgp5" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.824092 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.825021 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.831286 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.854643 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hjg5\" (UniqueName: \"kubernetes.io/projected/aa79cab5-ba24-4706-b428-8182c776848d-kube-api-access-6hjg5\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.856705 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-combined-ca-bundle\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.857399 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovn-rundir\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.857503 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-dns-svc\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.857530 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c521e9a3-0d64-4658-8a00-690d20e619d0-config\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.857585 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcfxw\" (UniqueName: \"kubernetes.io/projected/c521e9a3-0d64-4658-8a00-690d20e619d0-kube-api-access-hcfxw\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.857754 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.857878 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.858031 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.858188 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-config\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.858309 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovs-rundir\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.858745 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovn-rundir\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.861116 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovs-rundir\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.862447 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c521e9a3-0d64-4658-8a00-690d20e619d0-config\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.865563 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.867663 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.876607 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-combined-ca-bundle\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.891193 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcfxw\" (UniqueName: \"kubernetes.io/projected/c521e9a3-0d64-4658-8a00-690d20e619d0-kube-api-access-hcfxw\") pod \"ovn-controller-metrics-m2b6x\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.961812 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.961895 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-config\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.961966 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-config\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.962073 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hjg5\" (UniqueName: \"kubernetes.io/projected/aa79cab5-ba24-4706-b428-8182c776848d-kube-api-access-6hjg5\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.962157 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.962180 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z69d6\" (UniqueName: \"kubernetes.io/projected/2dfd2e16-b145-4464-8c5a-bff9a5908d45-kube-api-access-z69d6\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.962237 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.962342 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-scripts\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.963130 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-dns-svc\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.964642 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.964767 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.964805 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.963643 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.964021 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-dns-svc\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.964119 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-config\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.966399 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:43 crc kubenswrapper[4944]: I1124 09:09:43.978450 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hjg5\" (UniqueName: \"kubernetes.io/projected/aa79cab5-ba24-4706-b428-8182c776848d-kube-api-access-6hjg5\") pod \"dnsmasq-dns-8554648995-xsfw4\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.067432 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.067788 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z69d6\" (UniqueName: \"kubernetes.io/projected/2dfd2e16-b145-4464-8c5a-bff9a5908d45-kube-api-access-z69d6\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.067815 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.067966 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-scripts\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.068032 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.068076 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.068118 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-config\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.068982 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-config\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.069458 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.070103 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-scripts\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.072948 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.074080 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.075089 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.087395 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z69d6\" (UniqueName: \"kubernetes.io/projected/2dfd2e16-b145-4464-8c5a-bff9a5908d45-kube-api-access-z69d6\") pod \"ovn-northd-0\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.170993 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.199035 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.214506 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.304792 4944 generic.go:334] "Generic (PLEG): container finished" podID="d9e4a45c-7aa7-4c04-8c86-a7927a7ada80" containerID="751b73b489bf6597417b5d9527e9febd188bf9c2d66821855286a3d83547ec5d" exitCode=0 Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.305346 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0cdc-account-create-p4c9q" event={"ID":"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80","Type":"ContainerDied","Data":"751b73b489bf6597417b5d9527e9febd188bf9c2d66821855286a3d83547ec5d"} Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.305390 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0cdc-account-create-p4c9q" event={"ID":"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80","Type":"ContainerStarted","Data":"08d9bc032c492129324fe0b78694ba4d3ea10cf243055e137988cf9e86fd6d5f"} Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.333841 4944 generic.go:334] "Generic (PLEG): container finished" podID="6a5dfba8-039e-4bcd-ab5b-f654731ecfb4" containerID="3f8102c9e1131a4291d679c8ceadc0d3f525ccd386d483f44e28e313702eb963" exitCode=0 Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.333943 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d4j84" event={"ID":"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4","Type":"ContainerDied","Data":"3f8102c9e1131a4291d679c8ceadc0d3f525ccd386d483f44e28e313702eb963"} Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.333973 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d4j84" event={"ID":"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4","Type":"ContainerStarted","Data":"e30c3526deb38d1c70b92d85c56094c301c9d9c8716ae73f940281e947447cb5"} Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.336711 4944 generic.go:334] "Generic (PLEG): container finished" podID="1cc3188b-7842-4aa0-accd-cfa2e303f001" containerID="0fe726f95d087770f3a55da5b8318319bcd580d793904dd02e38159fed22e981" exitCode=0 Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.336767 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1ab7-account-create-c4ljx" event={"ID":"1cc3188b-7842-4aa0-accd-cfa2e303f001","Type":"ContainerDied","Data":"0fe726f95d087770f3a55da5b8318319bcd580d793904dd02e38159fed22e981"} Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.347825 4944 generic.go:334] "Generic (PLEG): container finished" podID="d07bba62-af8e-415e-a68d-ed9bf7602bd6" containerID="88dc28611f87d15ce858b43d3b87af57347bf25a3db8b6f40bb1b9316cec873b" exitCode=0 Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.348104 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" podUID="d14b3bac-81cb-4c72-a8bf-6bbfd829575f" containerName="dnsmasq-dns" containerID="cri-o://64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c" gracePeriod=10 Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.348478 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" podUID="67127c77-12a7-41bd-a975-863417818765" containerName="dnsmasq-dns" containerID="cri-o://deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697" gracePeriod=10 Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.348631 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wgxhq" event={"ID":"d07bba62-af8e-415e-a68d-ed9bf7602bd6","Type":"ContainerDied","Data":"88dc28611f87d15ce858b43d3b87af57347bf25a3db8b6f40bb1b9316cec873b"} Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.365819 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-srfqr"] Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.401662 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.458525 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9ct72"] Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.460066 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.571806 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9ct72"] Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.584412 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.584478 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-config\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.584554 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.584611 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.584635 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh9kq\" (UniqueName: \"kubernetes.io/projected/1330f372-b8dc-455e-9ef5-ddeb01f6df00-kube-api-access-nh9kq\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.595165 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-srfqr"] Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.686080 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.686172 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.686193 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh9kq\" (UniqueName: \"kubernetes.io/projected/1330f372-b8dc-455e-9ef5-ddeb01f6df00-kube-api-access-nh9kq\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.686231 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.686257 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-config\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.690614 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.690908 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-config\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.691662 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.692159 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.707334 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh9kq\" (UniqueName: \"kubernetes.io/projected/1330f372-b8dc-455e-9ef5-ddeb01f6df00-kube-api-access-nh9kq\") pod \"dnsmasq-dns-b8fbc5445-9ct72\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.722845 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.845827 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-m2b6x"] Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.876268 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 09:09:44 crc kubenswrapper[4944]: W1124 09:09:44.876366 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc521e9a3_0d64_4658_8a00_690d20e619d0.slice/crio-98bdfdff50d3326231df5a40e90f31540c340f817eb18f1f93ed08e357d3228f WatchSource:0}: Error finding container 98bdfdff50d3326231df5a40e90f31540c340f817eb18f1f93ed08e357d3228f: Status 404 returned error can't find the container with id 98bdfdff50d3326231df5a40e90f31540c340f817eb18f1f93ed08e357d3228f Nov 24 09:09:44 crc kubenswrapper[4944]: I1124 09:09:44.926032 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.095611 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-dns-svc\") pod \"67127c77-12a7-41bd-a975-863417818765\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.095989 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-config\") pod \"67127c77-12a7-41bd-a975-863417818765\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.096112 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnh4m\" (UniqueName: \"kubernetes.io/projected/67127c77-12a7-41bd-a975-863417818765-kube-api-access-nnh4m\") pod \"67127c77-12a7-41bd-a975-863417818765\" (UID: \"67127c77-12a7-41bd-a975-863417818765\") " Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.122281 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67127c77-12a7-41bd-a975-863417818765-kube-api-access-nnh4m" (OuterVolumeSpecName: "kube-api-access-nnh4m") pod "67127c77-12a7-41bd-a975-863417818765" (UID: "67127c77-12a7-41bd-a975-863417818765"). InnerVolumeSpecName "kube-api-access-nnh4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.174217 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "67127c77-12a7-41bd-a975-863417818765" (UID: "67127c77-12a7-41bd-a975-863417818765"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.202582 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.203614 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnh4m\" (UniqueName: \"kubernetes.io/projected/67127c77-12a7-41bd-a975-863417818765-kube-api-access-nnh4m\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.207065 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xsfw4"] Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.215293 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.264530 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-config" (OuterVolumeSpecName: "config") pod "67127c77-12a7-41bd-a975-863417818765" (UID: "67127c77-12a7-41bd-a975-863417818765"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.284622 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9ct72"] Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.304176 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-dns-svc\") pod \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.304291 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpdfj\" (UniqueName: \"kubernetes.io/projected/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-kube-api-access-mpdfj\") pod \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.304362 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-config\") pod \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\" (UID: \"d14b3bac-81cb-4c72-a8bf-6bbfd829575f\") " Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.304785 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67127c77-12a7-41bd-a975-863417818765-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.307340 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-kube-api-access-mpdfj" (OuterVolumeSpecName: "kube-api-access-mpdfj") pod "d14b3bac-81cb-4c72-a8bf-6bbfd829575f" (UID: "d14b3bac-81cb-4c72-a8bf-6bbfd829575f"). InnerVolumeSpecName "kube-api-access-mpdfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.362658 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-m2b6x" event={"ID":"c521e9a3-0d64-4658-8a00-690d20e619d0","Type":"ContainerStarted","Data":"98bdfdff50d3326231df5a40e90f31540c340f817eb18f1f93ed08e357d3228f"} Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.364000 4944 generic.go:334] "Generic (PLEG): container finished" podID="808a6dc9-1bea-4f98-8f9d-0d5101682ab6" containerID="64a1b581b646aff57dbc4ef2e211bdc9fb57397b6c9e853f4cd774cd066fd4d4" exitCode=0 Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.364066 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" event={"ID":"808a6dc9-1bea-4f98-8f9d-0d5101682ab6","Type":"ContainerDied","Data":"64a1b581b646aff57dbc4ef2e211bdc9fb57397b6c9e853f4cd774cd066fd4d4"} Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.364088 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" event={"ID":"808a6dc9-1bea-4f98-8f9d-0d5101682ab6","Type":"ContainerStarted","Data":"f56f3b1fa364e79a1a6f530bbe4de9ae497560eb76fd20e4dce88b67c6265a5e"} Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.373136 4944 generic.go:334] "Generic (PLEG): container finished" podID="67127c77-12a7-41bd-a975-863417818765" containerID="deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697" exitCode=0 Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.373329 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" event={"ID":"67127c77-12a7-41bd-a975-863417818765","Type":"ContainerDied","Data":"deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697"} Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.373605 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" event={"ID":"67127c77-12a7-41bd-a975-863417818765","Type":"ContainerDied","Data":"d73ec33895f8f79b2c825ef4b432eed1043b02ad68be662b8639bc9fee1e5a05"} Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.373628 4944 scope.go:117] "RemoveContainer" containerID="deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.373394 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5mj5n" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.379620 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2dfd2e16-b145-4464-8c5a-bff9a5908d45","Type":"ContainerStarted","Data":"80a682e82fe06c1e1fb85d732a2b9c78e72af4a7b72dd32e5964e1b65bae10e7"} Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.382263 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xsfw4" event={"ID":"aa79cab5-ba24-4706-b428-8182c776848d","Type":"ContainerStarted","Data":"889f782b87784080f80f198ddc61785ebf44170493e68e8b6ecd853ec0c7eacd"} Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.384176 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" event={"ID":"1330f372-b8dc-455e-9ef5-ddeb01f6df00","Type":"ContainerStarted","Data":"3d1320536f688fa86f8708aa1fbb55d9632f44fe22431ef74dedb3d03369572c"} Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.394192 4944 generic.go:334] "Generic (PLEG): container finished" podID="d14b3bac-81cb-4c72-a8bf-6bbfd829575f" containerID="64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c" exitCode=0 Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.394964 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.395192 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" event={"ID":"d14b3bac-81cb-4c72-a8bf-6bbfd829575f","Type":"ContainerDied","Data":"64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c"} Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.395231 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-6dkvq" event={"ID":"d14b3bac-81cb-4c72-a8bf-6bbfd829575f","Type":"ContainerDied","Data":"662d52bae705e0648d25f4b727ccb945185d738badecbb64047e24b2fc2dccb8"} Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.405932 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpdfj\" (UniqueName: \"kubernetes.io/projected/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-kube-api-access-mpdfj\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.420395 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d14b3bac-81cb-4c72-a8bf-6bbfd829575f" (UID: "d14b3bac-81cb-4c72-a8bf-6bbfd829575f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.424488 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-config" (OuterVolumeSpecName: "config") pod "d14b3bac-81cb-4c72-a8bf-6bbfd829575f" (UID: "d14b3bac-81cb-4c72-a8bf-6bbfd829575f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.446272 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5mj5n"] Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.452407 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5mj5n"] Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.493673 4944 scope.go:117] "RemoveContainer" containerID="434592ad9cf533bd4d47b291b28cce6adadd5b13a285800b305de9f00b8ecbaf" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.508397 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.508436 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d14b3bac-81cb-4c72-a8bf-6bbfd829575f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.560537 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 24 09:09:45 crc kubenswrapper[4944]: E1124 09:09:45.566092 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d14b3bac-81cb-4c72-a8bf-6bbfd829575f" containerName="init" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.566125 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d14b3bac-81cb-4c72-a8bf-6bbfd829575f" containerName="init" Nov 24 09:09:45 crc kubenswrapper[4944]: E1124 09:09:45.566159 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d14b3bac-81cb-4c72-a8bf-6bbfd829575f" containerName="dnsmasq-dns" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.566167 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d14b3bac-81cb-4c72-a8bf-6bbfd829575f" containerName="dnsmasq-dns" Nov 24 09:09:45 crc kubenswrapper[4944]: E1124 09:09:45.566195 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67127c77-12a7-41bd-a975-863417818765" containerName="dnsmasq-dns" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.566207 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="67127c77-12a7-41bd-a975-863417818765" containerName="dnsmasq-dns" Nov 24 09:09:45 crc kubenswrapper[4944]: E1124 09:09:45.566223 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67127c77-12a7-41bd-a975-863417818765" containerName="init" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.566232 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="67127c77-12a7-41bd-a975-863417818765" containerName="init" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.566709 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="67127c77-12a7-41bd-a975-863417818765" containerName="dnsmasq-dns" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.566756 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d14b3bac-81cb-4c72-a8bf-6bbfd829575f" containerName="dnsmasq-dns" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.594715 4944 scope.go:117] "RemoveContainer" containerID="deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697" Nov 24 09:09:45 crc kubenswrapper[4944]: E1124 09:09:45.595197 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697\": container with ID starting with deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697 not found: ID does not exist" containerID="deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.595237 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697"} err="failed to get container status \"deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697\": rpc error: code = NotFound desc = could not find container \"deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697\": container with ID starting with deec385bf174ada865986a1d9d32e2cf9129a0fb03b12757359ed66e3f2a3697 not found: ID does not exist" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.595265 4944 scope.go:117] "RemoveContainer" containerID="434592ad9cf533bd4d47b291b28cce6adadd5b13a285800b305de9f00b8ecbaf" Nov 24 09:09:45 crc kubenswrapper[4944]: E1124 09:09:45.595653 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"434592ad9cf533bd4d47b291b28cce6adadd5b13a285800b305de9f00b8ecbaf\": container with ID starting with 434592ad9cf533bd4d47b291b28cce6adadd5b13a285800b305de9f00b8ecbaf not found: ID does not exist" containerID="434592ad9cf533bd4d47b291b28cce6adadd5b13a285800b305de9f00b8ecbaf" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.595670 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"434592ad9cf533bd4d47b291b28cce6adadd5b13a285800b305de9f00b8ecbaf"} err="failed to get container status \"434592ad9cf533bd4d47b291b28cce6adadd5b13a285800b305de9f00b8ecbaf\": rpc error: code = NotFound desc = could not find container \"434592ad9cf533bd4d47b291b28cce6adadd5b13a285800b305de9f00b8ecbaf\": container with ID starting with 434592ad9cf533bd4d47b291b28cce6adadd5b13a285800b305de9f00b8ecbaf not found: ID does not exist" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.595683 4944 scope.go:117] "RemoveContainer" containerID="64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.602973 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.603174 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.606674 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.607073 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.607264 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.607525 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-r4kbp" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.701554 4944 scope.go:117] "RemoveContainer" containerID="24979742899bd5b53f8ded8da1e731c34e36522cfe9eb536f8b735d53e121eb6" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.713392 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-lock\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.713446 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.713533 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxs7n\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-kube-api-access-nxs7n\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.713575 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-cache\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.713594 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.762171 4944 scope.go:117] "RemoveContainer" containerID="64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c" Nov 24 09:09:45 crc kubenswrapper[4944]: E1124 09:09:45.765756 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c\": container with ID starting with 64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c not found: ID does not exist" containerID="64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.765830 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c"} err="failed to get container status \"64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c\": rpc error: code = NotFound desc = could not find container \"64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c\": container with ID starting with 64460950c7137c53f580142095f80b771c35966f3fee87c8ac1a3cae9090fd3c not found: ID does not exist" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.765872 4944 scope.go:117] "RemoveContainer" containerID="24979742899bd5b53f8ded8da1e731c34e36522cfe9eb536f8b735d53e121eb6" Nov 24 09:09:45 crc kubenswrapper[4944]: E1124 09:09:45.766334 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24979742899bd5b53f8ded8da1e731c34e36522cfe9eb536f8b735d53e121eb6\": container with ID starting with 24979742899bd5b53f8ded8da1e731c34e36522cfe9eb536f8b735d53e121eb6 not found: ID does not exist" containerID="24979742899bd5b53f8ded8da1e731c34e36522cfe9eb536f8b735d53e121eb6" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.766384 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24979742899bd5b53f8ded8da1e731c34e36522cfe9eb536f8b735d53e121eb6"} err="failed to get container status \"24979742899bd5b53f8ded8da1e731c34e36522cfe9eb536f8b735d53e121eb6\": rpc error: code = NotFound desc = could not find container \"24979742899bd5b53f8ded8da1e731c34e36522cfe9eb536f8b735d53e121eb6\": container with ID starting with 24979742899bd5b53f8ded8da1e731c34e36522cfe9eb536f8b735d53e121eb6 not found: ID does not exist" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.815461 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-cache\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.815508 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.815563 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-lock\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.815590 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.815684 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxs7n\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-kube-api-access-nxs7n\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.816707 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-cache\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.816980 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: E1124 09:09:45.823075 4944 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 09:09:45 crc kubenswrapper[4944]: E1124 09:09:45.823115 4944 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 09:09:45 crc kubenswrapper[4944]: E1124 09:09:45.823181 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift podName:6918e56a-00f4-4f99-b3ef-c65be06c428a nodeName:}" failed. No retries permitted until 2025-11-24 09:09:46.323158536 +0000 UTC m=+1046.857598998 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift") pod "swift-storage-0" (UID: "6918e56a-00f4-4f99-b3ef-c65be06c428a") : configmap "swift-ring-files" not found Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.823401 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-lock\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.857463 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.861783 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxs7n\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-kube-api-access-nxs7n\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.917065 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6dkvq"] Nov 24 09:09:45 crc kubenswrapper[4944]: I1124 09:09:45.950278 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6dkvq"] Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.047760 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0cdc-account-create-p4c9q" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.056606 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4j84" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.080393 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.086271 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wgxhq" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.096821 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1ab7-account-create-c4ljx" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.120487 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zf9p\" (UniqueName: \"kubernetes.io/projected/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-kube-api-access-5zf9p\") pod \"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80\" (UID: \"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.120559 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-operator-scripts\") pod \"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80\" (UID: \"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.120587 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mstp4\" (UniqueName: \"kubernetes.io/projected/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-kube-api-access-mstp4\") pod \"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4\" (UID: \"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.121316 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d9e4a45c-7aa7-4c04-8c86-a7927a7ada80" (UID: "d9e4a45c-7aa7-4c04-8c86-a7927a7ada80"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.121461 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-operator-scripts\") pod \"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4\" (UID: \"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.121964 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6a5dfba8-039e-4bcd-ab5b-f654731ecfb4" (UID: "6a5dfba8-039e-4bcd-ab5b-f654731ecfb4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.122709 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.122735 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.125028 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-kube-api-access-5zf9p" (OuterVolumeSpecName: "kube-api-access-5zf9p") pod "d9e4a45c-7aa7-4c04-8c86-a7927a7ada80" (UID: "d9e4a45c-7aa7-4c04-8c86-a7927a7ada80"). InnerVolumeSpecName "kube-api-access-5zf9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.125132 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-kube-api-access-mstp4" (OuterVolumeSpecName: "kube-api-access-mstp4") pod "6a5dfba8-039e-4bcd-ab5b-f654731ecfb4" (UID: "6a5dfba8-039e-4bcd-ab5b-f654731ecfb4"). InnerVolumeSpecName "kube-api-access-mstp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.224154 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-dns-svc\") pod \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.224326 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmghx\" (UniqueName: \"kubernetes.io/projected/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-kube-api-access-pmghx\") pod \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.224517 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qwkn\" (UniqueName: \"kubernetes.io/projected/d07bba62-af8e-415e-a68d-ed9bf7602bd6-kube-api-access-6qwkn\") pod \"d07bba62-af8e-415e-a68d-ed9bf7602bd6\" (UID: \"d07bba62-af8e-415e-a68d-ed9bf7602bd6\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.224594 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkrqv\" (UniqueName: \"kubernetes.io/projected/1cc3188b-7842-4aa0-accd-cfa2e303f001-kube-api-access-bkrqv\") pod \"1cc3188b-7842-4aa0-accd-cfa2e303f001\" (UID: \"1cc3188b-7842-4aa0-accd-cfa2e303f001\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.224635 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-config\") pod \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.224681 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d07bba62-af8e-415e-a68d-ed9bf7602bd6-operator-scripts\") pod \"d07bba62-af8e-415e-a68d-ed9bf7602bd6\" (UID: \"d07bba62-af8e-415e-a68d-ed9bf7602bd6\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.224733 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cc3188b-7842-4aa0-accd-cfa2e303f001-operator-scripts\") pod \"1cc3188b-7842-4aa0-accd-cfa2e303f001\" (UID: \"1cc3188b-7842-4aa0-accd-cfa2e303f001\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.224761 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-ovsdbserver-nb\") pod \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\" (UID: \"808a6dc9-1bea-4f98-8f9d-0d5101682ab6\") " Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.225459 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zf9p\" (UniqueName: \"kubernetes.io/projected/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80-kube-api-access-5zf9p\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.225493 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mstp4\" (UniqueName: \"kubernetes.io/projected/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4-kube-api-access-mstp4\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.229236 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cc3188b-7842-4aa0-accd-cfa2e303f001-kube-api-access-bkrqv" (OuterVolumeSpecName: "kube-api-access-bkrqv") pod "1cc3188b-7842-4aa0-accd-cfa2e303f001" (UID: "1cc3188b-7842-4aa0-accd-cfa2e303f001"). InnerVolumeSpecName "kube-api-access-bkrqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.229671 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cc3188b-7842-4aa0-accd-cfa2e303f001-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1cc3188b-7842-4aa0-accd-cfa2e303f001" (UID: "1cc3188b-7842-4aa0-accd-cfa2e303f001"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.229750 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d07bba62-af8e-415e-a68d-ed9bf7602bd6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d07bba62-af8e-415e-a68d-ed9bf7602bd6" (UID: "d07bba62-af8e-415e-a68d-ed9bf7602bd6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.230397 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-kube-api-access-pmghx" (OuterVolumeSpecName: "kube-api-access-pmghx") pod "808a6dc9-1bea-4f98-8f9d-0d5101682ab6" (UID: "808a6dc9-1bea-4f98-8f9d-0d5101682ab6"). InnerVolumeSpecName "kube-api-access-pmghx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.235509 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d07bba62-af8e-415e-a68d-ed9bf7602bd6-kube-api-access-6qwkn" (OuterVolumeSpecName: "kube-api-access-6qwkn") pod "d07bba62-af8e-415e-a68d-ed9bf7602bd6" (UID: "d07bba62-af8e-415e-a68d-ed9bf7602bd6"). InnerVolumeSpecName "kube-api-access-6qwkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.248085 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-config" (OuterVolumeSpecName: "config") pod "808a6dc9-1bea-4f98-8f9d-0d5101682ab6" (UID: "808a6dc9-1bea-4f98-8f9d-0d5101682ab6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.248318 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "808a6dc9-1bea-4f98-8f9d-0d5101682ab6" (UID: "808a6dc9-1bea-4f98-8f9d-0d5101682ab6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.255166 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "808a6dc9-1bea-4f98-8f9d-0d5101682ab6" (UID: "808a6dc9-1bea-4f98-8f9d-0d5101682ab6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.288922 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67127c77-12a7-41bd-a975-863417818765" path="/var/lib/kubelet/pods/67127c77-12a7-41bd-a975-863417818765/volumes" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.289683 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d14b3bac-81cb-4c72-a8bf-6bbfd829575f" path="/var/lib/kubelet/pods/d14b3bac-81cb-4c72-a8bf-6bbfd829575f/volumes" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.327071 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.327503 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkrqv\" (UniqueName: \"kubernetes.io/projected/1cc3188b-7842-4aa0-accd-cfa2e303f001-kube-api-access-bkrqv\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.327522 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.327536 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d07bba62-af8e-415e-a68d-ed9bf7602bd6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.327548 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cc3188b-7842-4aa0-accd-cfa2e303f001-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.327558 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.327569 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.327581 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmghx\" (UniqueName: \"kubernetes.io/projected/808a6dc9-1bea-4f98-8f9d-0d5101682ab6-kube-api-access-pmghx\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.327592 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qwkn\" (UniqueName: \"kubernetes.io/projected/d07bba62-af8e-415e-a68d-ed9bf7602bd6-kube-api-access-6qwkn\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:46 crc kubenswrapper[4944]: E1124 09:09:46.327553 4944 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 09:09:46 crc kubenswrapper[4944]: E1124 09:09:46.327622 4944 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 09:09:46 crc kubenswrapper[4944]: E1124 09:09:46.327710 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift podName:6918e56a-00f4-4f99-b3ef-c65be06c428a nodeName:}" failed. No retries permitted until 2025-11-24 09:09:47.327692357 +0000 UTC m=+1047.862132819 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift") pod "swift-storage-0" (UID: "6918e56a-00f4-4f99-b3ef-c65be06c428a") : configmap "swift-ring-files" not found Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.405333 4944 generic.go:334] "Generic (PLEG): container finished" podID="aa79cab5-ba24-4706-b428-8182c776848d" containerID="5d4af15a34d2472d4b3fd6ab99e360807933e36f8933db55a6f0707ef6e7fb04" exitCode=0 Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.405414 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xsfw4" event={"ID":"aa79cab5-ba24-4706-b428-8182c776848d","Type":"ContainerDied","Data":"5d4af15a34d2472d4b3fd6ab99e360807933e36f8933db55a6f0707ef6e7fb04"} Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.408463 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d4j84" event={"ID":"6a5dfba8-039e-4bcd-ab5b-f654731ecfb4","Type":"ContainerDied","Data":"e30c3526deb38d1c70b92d85c56094c301c9d9c8716ae73f940281e947447cb5"} Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.408505 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e30c3526deb38d1c70b92d85c56094c301c9d9c8716ae73f940281e947447cb5" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.408568 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4j84" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.419768 4944 generic.go:334] "Generic (PLEG): container finished" podID="1330f372-b8dc-455e-9ef5-ddeb01f6df00" containerID="67563f065abdc1b0605307ff47b7a0c04feebb212462f56b94c8960913c9012e" exitCode=0 Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.419873 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" event={"ID":"1330f372-b8dc-455e-9ef5-ddeb01f6df00","Type":"ContainerDied","Data":"67563f065abdc1b0605307ff47b7a0c04feebb212462f56b94c8960913c9012e"} Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.426236 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1ab7-account-create-c4ljx" event={"ID":"1cc3188b-7842-4aa0-accd-cfa2e303f001","Type":"ContainerDied","Data":"d858c06520b2d55277f3f86d5d9767488799c4dbcb8734327daa45899c4c3b32"} Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.426281 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d858c06520b2d55277f3f86d5d9767488799c4dbcb8734327daa45899c4c3b32" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.426345 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1ab7-account-create-c4ljx" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.444248 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" event={"ID":"808a6dc9-1bea-4f98-8f9d-0d5101682ab6","Type":"ContainerDied","Data":"f56f3b1fa364e79a1a6f530bbe4de9ae497560eb76fd20e4dce88b67c6265a5e"} Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.444293 4944 scope.go:117] "RemoveContainer" containerID="64a1b581b646aff57dbc4ef2e211bdc9fb57397b6c9e853f4cd774cd066fd4d4" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.444390 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-srfqr" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.452218 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-0cdc-account-create-p4c9q" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.452210 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-0cdc-account-create-p4c9q" event={"ID":"d9e4a45c-7aa7-4c04-8c86-a7927a7ada80","Type":"ContainerDied","Data":"08d9bc032c492129324fe0b78694ba4d3ea10cf243055e137988cf9e86fd6d5f"} Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.452339 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08d9bc032c492129324fe0b78694ba4d3ea10cf243055e137988cf9e86fd6d5f" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.454072 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-m2b6x" event={"ID":"c521e9a3-0d64-4658-8a00-690d20e619d0","Type":"ContainerStarted","Data":"b41e32588963f3e10325bdeeaf1f1bb9e18dd5fa9618690b49d86fb5f1d1c724"} Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.459514 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wgxhq" event={"ID":"d07bba62-af8e-415e-a68d-ed9bf7602bd6","Type":"ContainerDied","Data":"648ecb0fbdc7afb95f8e49dc906cd93b5361b8e0e256bcf322a72c08667e44e3"} Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.459559 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="648ecb0fbdc7afb95f8e49dc906cd93b5361b8e0e256bcf322a72c08667e44e3" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.464088 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wgxhq" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.490481 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-m2b6x" podStartSLOduration=3.490449024 podStartE2EDuration="3.490449024s" podCreationTimestamp="2025-11-24 09:09:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:09:46.478863495 +0000 UTC m=+1047.013303967" watchObservedRunningTime="2025-11-24 09:09:46.490449024 +0000 UTC m=+1047.024889486" Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.591570 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-srfqr"] Nov 24 09:09:46 crc kubenswrapper[4944]: I1124 09:09:46.624373 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-srfqr"] Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.356100 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:47 crc kubenswrapper[4944]: E1124 09:09:47.356318 4944 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 09:09:47 crc kubenswrapper[4944]: E1124 09:09:47.356782 4944 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 09:09:47 crc kubenswrapper[4944]: E1124 09:09:47.356840 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift podName:6918e56a-00f4-4f99-b3ef-c65be06c428a nodeName:}" failed. No retries permitted until 2025-11-24 09:09:49.356823306 +0000 UTC m=+1049.891263768 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift") pod "swift-storage-0" (UID: "6918e56a-00f4-4f99-b3ef-c65be06c428a") : configmap "swift-ring-files" not found Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.478173 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" event={"ID":"1330f372-b8dc-455e-9ef5-ddeb01f6df00","Type":"ContainerStarted","Data":"c54d394785155844676a6c1d096e0bb324d50ac7e9818cd6cbe354818c5132b5"} Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.478355 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.482025 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2dfd2e16-b145-4464-8c5a-bff9a5908d45","Type":"ContainerStarted","Data":"81af0f39619d4f0d4399fe088165e6c6fb343c03871f10888e1d787a30b8aa80"} Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.482098 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2dfd2e16-b145-4464-8c5a-bff9a5908d45","Type":"ContainerStarted","Data":"5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c"} Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.482217 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.485601 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xsfw4" event={"ID":"aa79cab5-ba24-4706-b428-8182c776848d","Type":"ContainerStarted","Data":"9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff"} Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.509728 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" podStartSLOduration=3.5097046389999997 podStartE2EDuration="3.509704639s" podCreationTimestamp="2025-11-24 09:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:09:47.505062941 +0000 UTC m=+1048.039503433" watchObservedRunningTime="2025-11-24 09:09:47.509704639 +0000 UTC m=+1048.044145091" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.528127 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.7151625040000003 podStartE2EDuration="4.528111505s" podCreationTimestamp="2025-11-24 09:09:43 +0000 UTC" firstStartedPulling="2025-11-24 09:09:44.945632519 +0000 UTC m=+1045.480072981" lastFinishedPulling="2025-11-24 09:09:46.75858152 +0000 UTC m=+1047.293021982" observedRunningTime="2025-11-24 09:09:47.525592765 +0000 UTC m=+1048.060033227" watchObservedRunningTime="2025-11-24 09:09:47.528111505 +0000 UTC m=+1048.062551967" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.549033 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-xsfw4" podStartSLOduration=4.549017062 podStartE2EDuration="4.549017062s" podCreationTimestamp="2025-11-24 09:09:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:09:47.547821414 +0000 UTC m=+1048.082261876" watchObservedRunningTime="2025-11-24 09:09:47.549017062 +0000 UTC m=+1048.083457524" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.996211 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-bgpkk"] Nov 24 09:09:47 crc kubenswrapper[4944]: E1124 09:09:47.996551 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cc3188b-7842-4aa0-accd-cfa2e303f001" containerName="mariadb-account-create" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.996568 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cc3188b-7842-4aa0-accd-cfa2e303f001" containerName="mariadb-account-create" Nov 24 09:09:47 crc kubenswrapper[4944]: E1124 09:09:47.996595 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d07bba62-af8e-415e-a68d-ed9bf7602bd6" containerName="mariadb-database-create" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.996601 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d07bba62-af8e-415e-a68d-ed9bf7602bd6" containerName="mariadb-database-create" Nov 24 09:09:47 crc kubenswrapper[4944]: E1124 09:09:47.996610 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="808a6dc9-1bea-4f98-8f9d-0d5101682ab6" containerName="init" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.996616 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="808a6dc9-1bea-4f98-8f9d-0d5101682ab6" containerName="init" Nov 24 09:09:47 crc kubenswrapper[4944]: E1124 09:09:47.996623 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9e4a45c-7aa7-4c04-8c86-a7927a7ada80" containerName="mariadb-account-create" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.996629 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9e4a45c-7aa7-4c04-8c86-a7927a7ada80" containerName="mariadb-account-create" Nov 24 09:09:47 crc kubenswrapper[4944]: E1124 09:09:47.996643 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5dfba8-039e-4bcd-ab5b-f654731ecfb4" containerName="mariadb-database-create" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.996648 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5dfba8-039e-4bcd-ab5b-f654731ecfb4" containerName="mariadb-database-create" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.996790 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cc3188b-7842-4aa0-accd-cfa2e303f001" containerName="mariadb-account-create" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.996803 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d07bba62-af8e-415e-a68d-ed9bf7602bd6" containerName="mariadb-database-create" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.996817 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5dfba8-039e-4bcd-ab5b-f654731ecfb4" containerName="mariadb-database-create" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.996827 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9e4a45c-7aa7-4c04-8c86-a7927a7ada80" containerName="mariadb-account-create" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.996836 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="808a6dc9-1bea-4f98-8f9d-0d5101682ab6" containerName="init" Nov 24 09:09:47 crc kubenswrapper[4944]: I1124 09:09:47.997423 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bgpkk" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.007725 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bgpkk"] Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.069197 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmprt\" (UniqueName: \"kubernetes.io/projected/1ef47237-852b-485d-954b-285d07f44af3-kube-api-access-lmprt\") pod \"glance-db-create-bgpkk\" (UID: \"1ef47237-852b-485d-954b-285d07f44af3\") " pod="openstack/glance-db-create-bgpkk" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.069257 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ef47237-852b-485d-954b-285d07f44af3-operator-scripts\") pod \"glance-db-create-bgpkk\" (UID: \"1ef47237-852b-485d-954b-285d07f44af3\") " pod="openstack/glance-db-create-bgpkk" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.104712 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-509f-account-create-6wsxt"] Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.105861 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-509f-account-create-6wsxt" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.108983 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.112638 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-509f-account-create-6wsxt"] Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.171341 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/787b579d-d704-469d-b0b9-c903c0eef0f9-operator-scripts\") pod \"glance-509f-account-create-6wsxt\" (UID: \"787b579d-d704-469d-b0b9-c903c0eef0f9\") " pod="openstack/glance-509f-account-create-6wsxt" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.171426 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4qfb\" (UniqueName: \"kubernetes.io/projected/787b579d-d704-469d-b0b9-c903c0eef0f9-kube-api-access-c4qfb\") pod \"glance-509f-account-create-6wsxt\" (UID: \"787b579d-d704-469d-b0b9-c903c0eef0f9\") " pod="openstack/glance-509f-account-create-6wsxt" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.171509 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmprt\" (UniqueName: \"kubernetes.io/projected/1ef47237-852b-485d-954b-285d07f44af3-kube-api-access-lmprt\") pod \"glance-db-create-bgpkk\" (UID: \"1ef47237-852b-485d-954b-285d07f44af3\") " pod="openstack/glance-db-create-bgpkk" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.171551 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ef47237-852b-485d-954b-285d07f44af3-operator-scripts\") pod \"glance-db-create-bgpkk\" (UID: \"1ef47237-852b-485d-954b-285d07f44af3\") " pod="openstack/glance-db-create-bgpkk" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.176506 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ef47237-852b-485d-954b-285d07f44af3-operator-scripts\") pod \"glance-db-create-bgpkk\" (UID: \"1ef47237-852b-485d-954b-285d07f44af3\") " pod="openstack/glance-db-create-bgpkk" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.189470 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmprt\" (UniqueName: \"kubernetes.io/projected/1ef47237-852b-485d-954b-285d07f44af3-kube-api-access-lmprt\") pod \"glance-db-create-bgpkk\" (UID: \"1ef47237-852b-485d-954b-285d07f44af3\") " pod="openstack/glance-db-create-bgpkk" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.273834 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/787b579d-d704-469d-b0b9-c903c0eef0f9-operator-scripts\") pod \"glance-509f-account-create-6wsxt\" (UID: \"787b579d-d704-469d-b0b9-c903c0eef0f9\") " pod="openstack/glance-509f-account-create-6wsxt" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.273898 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4qfb\" (UniqueName: \"kubernetes.io/projected/787b579d-d704-469d-b0b9-c903c0eef0f9-kube-api-access-c4qfb\") pod \"glance-509f-account-create-6wsxt\" (UID: \"787b579d-d704-469d-b0b9-c903c0eef0f9\") " pod="openstack/glance-509f-account-create-6wsxt" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.274555 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/787b579d-d704-469d-b0b9-c903c0eef0f9-operator-scripts\") pod \"glance-509f-account-create-6wsxt\" (UID: \"787b579d-d704-469d-b0b9-c903c0eef0f9\") " pod="openstack/glance-509f-account-create-6wsxt" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.286369 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="808a6dc9-1bea-4f98-8f9d-0d5101682ab6" path="/var/lib/kubelet/pods/808a6dc9-1bea-4f98-8f9d-0d5101682ab6/volumes" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.291701 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4qfb\" (UniqueName: \"kubernetes.io/projected/787b579d-d704-469d-b0b9-c903c0eef0f9-kube-api-access-c4qfb\") pod \"glance-509f-account-create-6wsxt\" (UID: \"787b579d-d704-469d-b0b9-c903c0eef0f9\") " pod="openstack/glance-509f-account-create-6wsxt" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.344057 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bgpkk" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.428089 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-509f-account-create-6wsxt" Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.493576 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:48 crc kubenswrapper[4944]: W1124 09:09:48.632711 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ef47237_852b_485d_954b_285d07f44af3.slice/crio-6ff4b19fbf82ea993147bf62f5201f20c837bc4681fe2191babdc450c27e96c1 WatchSource:0}: Error finding container 6ff4b19fbf82ea993147bf62f5201f20c837bc4681fe2191babdc450c27e96c1: Status 404 returned error can't find the container with id 6ff4b19fbf82ea993147bf62f5201f20c837bc4681fe2191babdc450c27e96c1 Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.640467 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bgpkk"] Nov 24 09:09:48 crc kubenswrapper[4944]: I1124 09:09:48.953298 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-509f-account-create-6wsxt"] Nov 24 09:09:48 crc kubenswrapper[4944]: W1124 09:09:48.963222 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod787b579d_d704_469d_b0b9_c903c0eef0f9.slice/crio-942c3465574fe66c797e0e42f8f61e18af41d22366c9456362c67fe4eef46e4f WatchSource:0}: Error finding container 942c3465574fe66c797e0e42f8f61e18af41d22366c9456362c67fe4eef46e4f: Status 404 returned error can't find the container with id 942c3465574fe66c797e0e42f8f61e18af41d22366c9456362c67fe4eef46e4f Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.392453 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:49 crc kubenswrapper[4944]: E1124 09:09:49.392975 4944 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 09:09:49 crc kubenswrapper[4944]: E1124 09:09:49.392995 4944 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 09:09:49 crc kubenswrapper[4944]: E1124 09:09:49.393057 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift podName:6918e56a-00f4-4f99-b3ef-c65be06c428a nodeName:}" failed. No retries permitted until 2025-11-24 09:09:53.393025553 +0000 UTC m=+1053.927466015 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift") pod "swift-storage-0" (UID: "6918e56a-00f4-4f99-b3ef-c65be06c428a") : configmap "swift-ring-files" not found Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.468238 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-mpksr"] Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.469302 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.473041 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.473869 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.477337 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.484638 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-mpksr"] Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.532532 4944 generic.go:334] "Generic (PLEG): container finished" podID="787b579d-d704-469d-b0b9-c903c0eef0f9" containerID="66e04adc7715ce4a2e86d8053addf377b2b9540de52cbc2b0ed756013704a659" exitCode=0 Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.532616 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-509f-account-create-6wsxt" event={"ID":"787b579d-d704-469d-b0b9-c903c0eef0f9","Type":"ContainerDied","Data":"66e04adc7715ce4a2e86d8053addf377b2b9540de52cbc2b0ed756013704a659"} Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.533034 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-509f-account-create-6wsxt" event={"ID":"787b579d-d704-469d-b0b9-c903c0eef0f9","Type":"ContainerStarted","Data":"942c3465574fe66c797e0e42f8f61e18af41d22366c9456362c67fe4eef46e4f"} Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.543725 4944 generic.go:334] "Generic (PLEG): container finished" podID="1ef47237-852b-485d-954b-285d07f44af3" containerID="99bf25824c378bb0feb5d33ed10f3a48b6d21f932f260473ccb0a6b7c666246d" exitCode=0 Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.543791 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bgpkk" event={"ID":"1ef47237-852b-485d-954b-285d07f44af3","Type":"ContainerDied","Data":"99bf25824c378bb0feb5d33ed10f3a48b6d21f932f260473ccb0a6b7c666246d"} Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.543821 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bgpkk" event={"ID":"1ef47237-852b-485d-954b-285d07f44af3","Type":"ContainerStarted","Data":"6ff4b19fbf82ea993147bf62f5201f20c837bc4681fe2191babdc450c27e96c1"} Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.605663 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5349072-d740-47d5-a667-05670688b5c5-etc-swift\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.605757 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-dispersionconf\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.605786 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-scripts\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.605805 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-ring-data-devices\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.605823 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5ll6\" (UniqueName: \"kubernetes.io/projected/e5349072-d740-47d5-a667-05670688b5c5-kube-api-access-v5ll6\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.605840 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-combined-ca-bundle\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.606022 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-swiftconf\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.707501 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-dispersionconf\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.707560 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-scripts\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.707579 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-ring-data-devices\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.707597 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5ll6\" (UniqueName: \"kubernetes.io/projected/e5349072-d740-47d5-a667-05670688b5c5-kube-api-access-v5ll6\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.707616 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-combined-ca-bundle\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.707661 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-swiftconf\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.707713 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5349072-d740-47d5-a667-05670688b5c5-etc-swift\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.708280 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-scripts\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.708527 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-ring-data-devices\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.708558 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5349072-d740-47d5-a667-05670688b5c5-etc-swift\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.713326 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-dispersionconf\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.713619 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-swiftconf\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.715750 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-combined-ca-bundle\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.723286 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5ll6\" (UniqueName: \"kubernetes.io/projected/e5349072-d740-47d5-a667-05670688b5c5-kube-api-access-v5ll6\") pod \"swift-ring-rebalance-mpksr\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:49 crc kubenswrapper[4944]: I1124 09:09:49.838317 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:09:50 crc kubenswrapper[4944]: I1124 09:09:50.332025 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-mpksr"] Nov 24 09:09:50 crc kubenswrapper[4944]: W1124 09:09:50.341490 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5349072_d740_47d5_a667_05670688b5c5.slice/crio-f880a4f7aae3e1574d61e303d2605cdc5f2983f9623366f3059b46d789a4aa73 WatchSource:0}: Error finding container f880a4f7aae3e1574d61e303d2605cdc5f2983f9623366f3059b46d789a4aa73: Status 404 returned error can't find the container with id f880a4f7aae3e1574d61e303d2605cdc5f2983f9623366f3059b46d789a4aa73 Nov 24 09:09:50 crc kubenswrapper[4944]: I1124 09:09:50.552472 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mpksr" event={"ID":"e5349072-d740-47d5-a667-05670688b5c5","Type":"ContainerStarted","Data":"f880a4f7aae3e1574d61e303d2605cdc5f2983f9623366f3059b46d789a4aa73"} Nov 24 09:09:50 crc kubenswrapper[4944]: I1124 09:09:50.977083 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-509f-account-create-6wsxt" Nov 24 09:09:50 crc kubenswrapper[4944]: I1124 09:09:50.984334 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bgpkk" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.136245 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmprt\" (UniqueName: \"kubernetes.io/projected/1ef47237-852b-485d-954b-285d07f44af3-kube-api-access-lmprt\") pod \"1ef47237-852b-485d-954b-285d07f44af3\" (UID: \"1ef47237-852b-485d-954b-285d07f44af3\") " Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.136301 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ef47237-852b-485d-954b-285d07f44af3-operator-scripts\") pod \"1ef47237-852b-485d-954b-285d07f44af3\" (UID: \"1ef47237-852b-485d-954b-285d07f44af3\") " Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.136363 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4qfb\" (UniqueName: \"kubernetes.io/projected/787b579d-d704-469d-b0b9-c903c0eef0f9-kube-api-access-c4qfb\") pod \"787b579d-d704-469d-b0b9-c903c0eef0f9\" (UID: \"787b579d-d704-469d-b0b9-c903c0eef0f9\") " Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.136451 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/787b579d-d704-469d-b0b9-c903c0eef0f9-operator-scripts\") pod \"787b579d-d704-469d-b0b9-c903c0eef0f9\" (UID: \"787b579d-d704-469d-b0b9-c903c0eef0f9\") " Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.137179 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/787b579d-d704-469d-b0b9-c903c0eef0f9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "787b579d-d704-469d-b0b9-c903c0eef0f9" (UID: "787b579d-d704-469d-b0b9-c903c0eef0f9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.137192 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ef47237-852b-485d-954b-285d07f44af3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ef47237-852b-485d-954b-285d07f44af3" (UID: "1ef47237-852b-485d-954b-285d07f44af3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.142344 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ef47237-852b-485d-954b-285d07f44af3-kube-api-access-lmprt" (OuterVolumeSpecName: "kube-api-access-lmprt") pod "1ef47237-852b-485d-954b-285d07f44af3" (UID: "1ef47237-852b-485d-954b-285d07f44af3"). InnerVolumeSpecName "kube-api-access-lmprt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.142571 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/787b579d-d704-469d-b0b9-c903c0eef0f9-kube-api-access-c4qfb" (OuterVolumeSpecName: "kube-api-access-c4qfb") pod "787b579d-d704-469d-b0b9-c903c0eef0f9" (UID: "787b579d-d704-469d-b0b9-c903c0eef0f9"). InnerVolumeSpecName "kube-api-access-c4qfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.239257 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmprt\" (UniqueName: \"kubernetes.io/projected/1ef47237-852b-485d-954b-285d07f44af3-kube-api-access-lmprt\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.239321 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ef47237-852b-485d-954b-285d07f44af3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.239335 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4qfb\" (UniqueName: \"kubernetes.io/projected/787b579d-d704-469d-b0b9-c903c0eef0f9-kube-api-access-c4qfb\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.239346 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/787b579d-d704-469d-b0b9-c903c0eef0f9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.562934 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-509f-account-create-6wsxt" event={"ID":"787b579d-d704-469d-b0b9-c903c0eef0f9","Type":"ContainerDied","Data":"942c3465574fe66c797e0e42f8f61e18af41d22366c9456362c67fe4eef46e4f"} Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.562970 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="942c3465574fe66c797e0e42f8f61e18af41d22366c9456362c67fe4eef46e4f" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.563018 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-509f-account-create-6wsxt" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.569144 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bgpkk" event={"ID":"1ef47237-852b-485d-954b-285d07f44af3","Type":"ContainerDied","Data":"6ff4b19fbf82ea993147bf62f5201f20c837bc4681fe2191babdc450c27e96c1"} Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.569198 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ff4b19fbf82ea993147bf62f5201f20c837bc4681fe2191babdc450c27e96c1" Nov 24 09:09:51 crc kubenswrapper[4944]: I1124 09:09:51.569156 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bgpkk" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.242975 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-vlctx"] Nov 24 09:09:53 crc kubenswrapper[4944]: E1124 09:09:53.243641 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef47237-852b-485d-954b-285d07f44af3" containerName="mariadb-database-create" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.243654 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef47237-852b-485d-954b-285d07f44af3" containerName="mariadb-database-create" Nov 24 09:09:53 crc kubenswrapper[4944]: E1124 09:09:53.243682 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="787b579d-d704-469d-b0b9-c903c0eef0f9" containerName="mariadb-account-create" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.243688 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="787b579d-d704-469d-b0b9-c903c0eef0f9" containerName="mariadb-account-create" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.243832 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="787b579d-d704-469d-b0b9-c903c0eef0f9" containerName="mariadb-account-create" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.243853 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ef47237-852b-485d-954b-285d07f44af3" containerName="mariadb-database-create" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.244450 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.246745 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.248804 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-9rvqc" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.264025 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-vlctx"] Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.388167 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-config-data\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.388263 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-db-sync-config-data\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.388292 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-combined-ca-bundle\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.388349 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zf98\" (UniqueName: \"kubernetes.io/projected/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-kube-api-access-7zf98\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.493592 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zf98\" (UniqueName: \"kubernetes.io/projected/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-kube-api-access-7zf98\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.493674 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.493708 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-config-data\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.493738 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-combined-ca-bundle\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.493756 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-db-sync-config-data\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: E1124 09:09:53.494749 4944 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 09:09:53 crc kubenswrapper[4944]: E1124 09:09:53.494767 4944 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 09:09:53 crc kubenswrapper[4944]: E1124 09:09:53.494807 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift podName:6918e56a-00f4-4f99-b3ef-c65be06c428a nodeName:}" failed. No retries permitted until 2025-11-24 09:10:01.494792683 +0000 UTC m=+1062.029233145 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift") pod "swift-storage-0" (UID: "6918e56a-00f4-4f99-b3ef-c65be06c428a") : configmap "swift-ring-files" not found Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.502608 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-db-sync-config-data\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.504293 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-config-data\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.516838 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-combined-ca-bundle\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.530617 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zf98\" (UniqueName: \"kubernetes.io/projected/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-kube-api-access-7zf98\") pod \"glance-db-sync-vlctx\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:53 crc kubenswrapper[4944]: I1124 09:09:53.568444 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vlctx" Nov 24 09:09:54 crc kubenswrapper[4944]: I1124 09:09:54.144940 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-vlctx"] Nov 24 09:09:54 crc kubenswrapper[4944]: W1124 09:09:54.154556 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4e37c3c_32ab_45ff_8b5a_a46fa5fe0d93.slice/crio-d56c29e4dbfc201dcde60401bdd060f3a7016b7ea7bbb29b97fdb2ef12644118 WatchSource:0}: Error finding container d56c29e4dbfc201dcde60401bdd060f3a7016b7ea7bbb29b97fdb2ef12644118: Status 404 returned error can't find the container with id d56c29e4dbfc201dcde60401bdd060f3a7016b7ea7bbb29b97fdb2ef12644118 Nov 24 09:09:54 crc kubenswrapper[4944]: I1124 09:09:54.201254 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:54 crc kubenswrapper[4944]: I1124 09:09:54.612893 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mpksr" event={"ID":"e5349072-d740-47d5-a667-05670688b5c5","Type":"ContainerStarted","Data":"86988b92220b8f00057d839eacd2728b2c903a679a7f8d2f627e8ac04be99bcd"} Nov 24 09:09:54 crc kubenswrapper[4944]: I1124 09:09:54.614412 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vlctx" event={"ID":"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93","Type":"ContainerStarted","Data":"d56c29e4dbfc201dcde60401bdd060f3a7016b7ea7bbb29b97fdb2ef12644118"} Nov 24 09:09:54 crc kubenswrapper[4944]: I1124 09:09:54.632664 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-mpksr" podStartSLOduration=2.526574421 podStartE2EDuration="5.632644367s" podCreationTimestamp="2025-11-24 09:09:49 +0000 UTC" firstStartedPulling="2025-11-24 09:09:50.343832446 +0000 UTC m=+1050.878272908" lastFinishedPulling="2025-11-24 09:09:53.449902402 +0000 UTC m=+1053.984342854" observedRunningTime="2025-11-24 09:09:54.632302287 +0000 UTC m=+1055.166742749" watchObservedRunningTime="2025-11-24 09:09:54.632644367 +0000 UTC m=+1055.167084829" Nov 24 09:09:54 crc kubenswrapper[4944]: I1124 09:09:54.724554 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:09:54 crc kubenswrapper[4944]: I1124 09:09:54.778451 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xsfw4"] Nov 24 09:09:54 crc kubenswrapper[4944]: I1124 09:09:54.778702 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-xsfw4" podUID="aa79cab5-ba24-4706-b428-8182c776848d" containerName="dnsmasq-dns" containerID="cri-o://9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff" gracePeriod=10 Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.214683 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.246114 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-nb\") pod \"aa79cab5-ba24-4706-b428-8182c776848d\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.246224 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-dns-svc\") pod \"aa79cab5-ba24-4706-b428-8182c776848d\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.246312 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-sb\") pod \"aa79cab5-ba24-4706-b428-8182c776848d\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.246383 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-config\") pod \"aa79cab5-ba24-4706-b428-8182c776848d\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.246463 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hjg5\" (UniqueName: \"kubernetes.io/projected/aa79cab5-ba24-4706-b428-8182c776848d-kube-api-access-6hjg5\") pod \"aa79cab5-ba24-4706-b428-8182c776848d\" (UID: \"aa79cab5-ba24-4706-b428-8182c776848d\") " Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.254039 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa79cab5-ba24-4706-b428-8182c776848d-kube-api-access-6hjg5" (OuterVolumeSpecName: "kube-api-access-6hjg5") pod "aa79cab5-ba24-4706-b428-8182c776848d" (UID: "aa79cab5-ba24-4706-b428-8182c776848d"). InnerVolumeSpecName "kube-api-access-6hjg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.289460 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa79cab5-ba24-4706-b428-8182c776848d" (UID: "aa79cab5-ba24-4706-b428-8182c776848d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.295326 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-config" (OuterVolumeSpecName: "config") pod "aa79cab5-ba24-4706-b428-8182c776848d" (UID: "aa79cab5-ba24-4706-b428-8182c776848d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.306256 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa79cab5-ba24-4706-b428-8182c776848d" (UID: "aa79cab5-ba24-4706-b428-8182c776848d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.308174 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa79cab5-ba24-4706-b428-8182c776848d" (UID: "aa79cab5-ba24-4706-b428-8182c776848d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.347723 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.347750 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hjg5\" (UniqueName: \"kubernetes.io/projected/aa79cab5-ba24-4706-b428-8182c776848d-kube-api-access-6hjg5\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.347759 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.347768 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.347775 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa79cab5-ba24-4706-b428-8182c776848d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.628726 4944 generic.go:334] "Generic (PLEG): container finished" podID="aa79cab5-ba24-4706-b428-8182c776848d" containerID="9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff" exitCode=0 Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.628815 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xsfw4" event={"ID":"aa79cab5-ba24-4706-b428-8182c776848d","Type":"ContainerDied","Data":"9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff"} Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.628885 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xsfw4" event={"ID":"aa79cab5-ba24-4706-b428-8182c776848d","Type":"ContainerDied","Data":"889f782b87784080f80f198ddc61785ebf44170493e68e8b6ecd853ec0c7eacd"} Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.628907 4944 scope.go:117] "RemoveContainer" containerID="9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.628958 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xsfw4" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.673999 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xsfw4"] Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.674967 4944 scope.go:117] "RemoveContainer" containerID="5d4af15a34d2472d4b3fd6ab99e360807933e36f8933db55a6f0707ef6e7fb04" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.680994 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xsfw4"] Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.707339 4944 scope.go:117] "RemoveContainer" containerID="9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff" Nov 24 09:09:55 crc kubenswrapper[4944]: E1124 09:09:55.707897 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff\": container with ID starting with 9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff not found: ID does not exist" containerID="9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.707967 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff"} err="failed to get container status \"9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff\": rpc error: code = NotFound desc = could not find container \"9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff\": container with ID starting with 9f38887168113331bedd2c3c9a053fdc662ff9df85f128b4f2a8f941e5793eff not found: ID does not exist" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.707996 4944 scope.go:117] "RemoveContainer" containerID="5d4af15a34d2472d4b3fd6ab99e360807933e36f8933db55a6f0707ef6e7fb04" Nov 24 09:09:55 crc kubenswrapper[4944]: E1124 09:09:55.708677 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d4af15a34d2472d4b3fd6ab99e360807933e36f8933db55a6f0707ef6e7fb04\": container with ID starting with 5d4af15a34d2472d4b3fd6ab99e360807933e36f8933db55a6f0707ef6e7fb04 not found: ID does not exist" containerID="5d4af15a34d2472d4b3fd6ab99e360807933e36f8933db55a6f0707ef6e7fb04" Nov 24 09:09:55 crc kubenswrapper[4944]: I1124 09:09:55.709493 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d4af15a34d2472d4b3fd6ab99e360807933e36f8933db55a6f0707ef6e7fb04"} err="failed to get container status \"5d4af15a34d2472d4b3fd6ab99e360807933e36f8933db55a6f0707ef6e7fb04\": rpc error: code = NotFound desc = could not find container \"5d4af15a34d2472d4b3fd6ab99e360807933e36f8933db55a6f0707ef6e7fb04\": container with ID starting with 5d4af15a34d2472d4b3fd6ab99e360807933e36f8933db55a6f0707ef6e7fb04 not found: ID does not exist" Nov 24 09:09:56 crc kubenswrapper[4944]: I1124 09:09:56.286538 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa79cab5-ba24-4706-b428-8182c776848d" path="/var/lib/kubelet/pods/aa79cab5-ba24-4706-b428-8182c776848d/volumes" Nov 24 09:09:59 crc kubenswrapper[4944]: I1124 09:09:59.285933 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 09:10:01 crc kubenswrapper[4944]: I1124 09:10:01.552443 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:10:01 crc kubenswrapper[4944]: I1124 09:10:01.568295 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift\") pod \"swift-storage-0\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " pod="openstack/swift-storage-0" Nov 24 09:10:01 crc kubenswrapper[4944]: I1124 09:10:01.697594 4944 generic.go:334] "Generic (PLEG): container finished" podID="e5349072-d740-47d5-a667-05670688b5c5" containerID="86988b92220b8f00057d839eacd2728b2c903a679a7f8d2f627e8ac04be99bcd" exitCode=0 Nov 24 09:10:01 crc kubenswrapper[4944]: I1124 09:10:01.697643 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mpksr" event={"ID":"e5349072-d740-47d5-a667-05670688b5c5","Type":"ContainerDied","Data":"86988b92220b8f00057d839eacd2728b2c903a679a7f8d2f627e8ac04be99bcd"} Nov 24 09:10:01 crc kubenswrapper[4944]: I1124 09:10:01.795605 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 09:10:04 crc kubenswrapper[4944]: I1124 09:10:04.721731 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mpksr" event={"ID":"e5349072-d740-47d5-a667-05670688b5c5","Type":"ContainerDied","Data":"f880a4f7aae3e1574d61e303d2605cdc5f2983f9623366f3059b46d789a4aa73"} Nov 24 09:10:04 crc kubenswrapper[4944]: I1124 09:10:04.722236 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f880a4f7aae3e1574d61e303d2605cdc5f2983f9623366f3059b46d789a4aa73" Nov 24 09:10:04 crc kubenswrapper[4944]: I1124 09:10:04.849469 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.002020 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5ll6\" (UniqueName: \"kubernetes.io/projected/e5349072-d740-47d5-a667-05670688b5c5-kube-api-access-v5ll6\") pod \"e5349072-d740-47d5-a667-05670688b5c5\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.002120 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-swiftconf\") pod \"e5349072-d740-47d5-a667-05670688b5c5\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.002151 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-scripts\") pod \"e5349072-d740-47d5-a667-05670688b5c5\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.002176 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-ring-data-devices\") pod \"e5349072-d740-47d5-a667-05670688b5c5\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.002269 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-dispersionconf\") pod \"e5349072-d740-47d5-a667-05670688b5c5\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.002309 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-combined-ca-bundle\") pod \"e5349072-d740-47d5-a667-05670688b5c5\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.002415 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5349072-d740-47d5-a667-05670688b5c5-etc-swift\") pod \"e5349072-d740-47d5-a667-05670688b5c5\" (UID: \"e5349072-d740-47d5-a667-05670688b5c5\") " Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.003283 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e5349072-d740-47d5-a667-05670688b5c5" (UID: "e5349072-d740-47d5-a667-05670688b5c5"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.003483 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5349072-d740-47d5-a667-05670688b5c5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e5349072-d740-47d5-a667-05670688b5c5" (UID: "e5349072-d740-47d5-a667-05670688b5c5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.012810 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5349072-d740-47d5-a667-05670688b5c5-kube-api-access-v5ll6" (OuterVolumeSpecName: "kube-api-access-v5ll6") pod "e5349072-d740-47d5-a667-05670688b5c5" (UID: "e5349072-d740-47d5-a667-05670688b5c5"). InnerVolumeSpecName "kube-api-access-v5ll6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.028914 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-scripts" (OuterVolumeSpecName: "scripts") pod "e5349072-d740-47d5-a667-05670688b5c5" (UID: "e5349072-d740-47d5-a667-05670688b5c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.033404 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5349072-d740-47d5-a667-05670688b5c5" (UID: "e5349072-d740-47d5-a667-05670688b5c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.034831 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e5349072-d740-47d5-a667-05670688b5c5" (UID: "e5349072-d740-47d5-a667-05670688b5c5"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.036491 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e5349072-d740-47d5-a667-05670688b5c5" (UID: "e5349072-d740-47d5-a667-05670688b5c5"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.104499 4944 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e5349072-d740-47d5-a667-05670688b5c5-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.104537 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5ll6\" (UniqueName: \"kubernetes.io/projected/e5349072-d740-47d5-a667-05670688b5c5-kube-api-access-v5ll6\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.104553 4944 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.104564 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.104576 4944 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e5349072-d740-47d5-a667-05670688b5c5-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.104587 4944 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.104656 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5349072-d740-47d5-a667-05670688b5c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.291400 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 09:10:05 crc kubenswrapper[4944]: W1124 09:10:05.314808 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6918e56a_00f4_4f99_b3ef_c65be06c428a.slice/crio-f531019f16c6b75dbbdfa316451f4b78f6ee7377c8c252e084ad48887aa300de WatchSource:0}: Error finding container f531019f16c6b75dbbdfa316451f4b78f6ee7377c8c252e084ad48887aa300de: Status 404 returned error can't find the container with id f531019f16c6b75dbbdfa316451f4b78f6ee7377c8c252e084ad48887aa300de Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.731721 4944 generic.go:334] "Generic (PLEG): container finished" podID="67e92d3f-3532-467f-bf38-c9c3107d4fcb" containerID="49cd4f1b400eddca6bdd6df980eb6336eb7f945c92b13a7f0c34e97beb31f078" exitCode=0 Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.731968 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"67e92d3f-3532-467f-bf38-c9c3107d4fcb","Type":"ContainerDied","Data":"49cd4f1b400eddca6bdd6df980eb6336eb7f945c92b13a7f0c34e97beb31f078"} Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.735079 4944 generic.go:334] "Generic (PLEG): container finished" podID="12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" containerID="0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033" exitCode=0 Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.735190 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5","Type":"ContainerDied","Data":"0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033"} Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.742619 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"f531019f16c6b75dbbdfa316451f4b78f6ee7377c8c252e084ad48887aa300de"} Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.745290 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vlctx" event={"ID":"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93","Type":"ContainerStarted","Data":"5f7f01bba3bbc83c216d9707bc817704d3d12a1909dfb8fada43a600b9229191"} Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.745311 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mpksr" Nov 24 09:10:05 crc kubenswrapper[4944]: I1124 09:10:05.778618 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-vlctx" podStartSLOduration=2.149642044 podStartE2EDuration="12.778595887s" podCreationTimestamp="2025-11-24 09:09:53 +0000 UTC" firstStartedPulling="2025-11-24 09:09:54.157106561 +0000 UTC m=+1054.691547023" lastFinishedPulling="2025-11-24 09:10:04.786060404 +0000 UTC m=+1065.320500866" observedRunningTime="2025-11-24 09:10:05.771800141 +0000 UTC m=+1066.306240603" watchObservedRunningTime="2025-11-24 09:10:05.778595887 +0000 UTC m=+1066.313036359" Nov 24 09:10:06 crc kubenswrapper[4944]: I1124 09:10:06.769245 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"67e92d3f-3532-467f-bf38-c9c3107d4fcb","Type":"ContainerStarted","Data":"ea9e352d0f24eca87b4c56a85b5935690423ccc0712e37d2ea78fde15ddd4336"} Nov 24 09:10:06 crc kubenswrapper[4944]: I1124 09:10:06.769922 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 09:10:06 crc kubenswrapper[4944]: I1124 09:10:06.773885 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5","Type":"ContainerStarted","Data":"07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256"} Nov 24 09:10:06 crc kubenswrapper[4944]: I1124 09:10:06.774800 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:10:06 crc kubenswrapper[4944]: I1124 09:10:06.776687 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"2842b8bb44b04f4fb4de40704e3d22283e14c59ce261e4a211bef21dbb6e7a96"} Nov 24 09:10:06 crc kubenswrapper[4944]: I1124 09:10:06.776721 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"3d31ceee24450aa9fe538636500f7868bbd37f5da8e0012f05ca3b457f8720eb"} Nov 24 09:10:06 crc kubenswrapper[4944]: I1124 09:10:06.799892 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.334087355 podStartE2EDuration="58.799875447s" podCreationTimestamp="2025-11-24 09:09:08 +0000 UTC" firstStartedPulling="2025-11-24 09:09:10.431695977 +0000 UTC m=+1010.966136439" lastFinishedPulling="2025-11-24 09:09:28.897484069 +0000 UTC m=+1029.431924531" observedRunningTime="2025-11-24 09:10:06.796663605 +0000 UTC m=+1067.331104067" watchObservedRunningTime="2025-11-24 09:10:06.799875447 +0000 UTC m=+1067.334315909" Nov 24 09:10:06 crc kubenswrapper[4944]: I1124 09:10:06.830643 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.823143611 podStartE2EDuration="59.830625147s" podCreationTimestamp="2025-11-24 09:09:07 +0000 UTC" firstStartedPulling="2025-11-24 09:09:09.886884203 +0000 UTC m=+1010.421324665" lastFinishedPulling="2025-11-24 09:09:28.894365739 +0000 UTC m=+1029.428806201" observedRunningTime="2025-11-24 09:10:06.827700704 +0000 UTC m=+1067.362141176" watchObservedRunningTime="2025-11-24 09:10:06.830625147 +0000 UTC m=+1067.365065609" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.335810 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-npmfc" podUID="69084f7b-b347-4fdf-917e-2f534f3cc47c" containerName="ovn-controller" probeResult="failure" output=< Nov 24 09:10:07 crc kubenswrapper[4944]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 09:10:07 crc kubenswrapper[4944]: > Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.339136 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.358057 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.562124 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-npmfc-config-zcbpb"] Nov 24 09:10:07 crc kubenswrapper[4944]: E1124 09:10:07.562698 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa79cab5-ba24-4706-b428-8182c776848d" containerName="dnsmasq-dns" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.562716 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa79cab5-ba24-4706-b428-8182c776848d" containerName="dnsmasq-dns" Nov 24 09:10:07 crc kubenswrapper[4944]: E1124 09:10:07.562741 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa79cab5-ba24-4706-b428-8182c776848d" containerName="init" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.562748 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa79cab5-ba24-4706-b428-8182c776848d" containerName="init" Nov 24 09:10:07 crc kubenswrapper[4944]: E1124 09:10:07.562765 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5349072-d740-47d5-a667-05670688b5c5" containerName="swift-ring-rebalance" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.562771 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5349072-d740-47d5-a667-05670688b5c5" containerName="swift-ring-rebalance" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.562915 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa79cab5-ba24-4706-b428-8182c776848d" containerName="dnsmasq-dns" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.562931 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5349072-d740-47d5-a667-05670688b5c5" containerName="swift-ring-rebalance" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.563485 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.566907 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.575575 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-npmfc-config-zcbpb"] Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.649182 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-scripts\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.649226 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpzvg\" (UniqueName: \"kubernetes.io/projected/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-kube-api-access-tpzvg\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.649249 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-log-ovn\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.649275 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-additional-scripts\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.649330 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.649382 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run-ovn\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.750912 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-scripts\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.750973 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpzvg\" (UniqueName: \"kubernetes.io/projected/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-kube-api-access-tpzvg\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.750999 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-log-ovn\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.751041 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-additional-scripts\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.751133 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.751190 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run-ovn\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.751405 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-log-ovn\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.751481 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run-ovn\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.751664 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.751970 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-additional-scripts\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.757995 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-scripts\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.770023 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpzvg\" (UniqueName: \"kubernetes.io/projected/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-kube-api-access-tpzvg\") pod \"ovn-controller-npmfc-config-zcbpb\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.789506 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"87e895d554127f520ab155e47e47d2faa5077bd4196802ecc68ee9b04403fe3d"} Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.789554 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"9311cf4164c78aa256b1228bc545098e231c1036813339b9399c8c0ccdf8463f"} Nov 24 09:10:07 crc kubenswrapper[4944]: I1124 09:10:07.892299 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:08 crc kubenswrapper[4944]: I1124 09:10:08.172298 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-npmfc-config-zcbpb"] Nov 24 09:10:08 crc kubenswrapper[4944]: I1124 09:10:08.811671 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-npmfc-config-zcbpb" event={"ID":"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4","Type":"ContainerStarted","Data":"78f08b95c177bc88f756e1a35fe2247319880a774db5ee06e28e464c5c5d83c2"} Nov 24 09:10:09 crc kubenswrapper[4944]: I1124 09:10:09.827748 4944 generic.go:334] "Generic (PLEG): container finished" podID="302a4ebf-b0a0-408c-8eb9-93009e2d9ec4" containerID="de79cb55674ca0207fc7a59b27c802ecc027ac030257f5a33d496e3c2b02b09a" exitCode=0 Nov 24 09:10:09 crc kubenswrapper[4944]: I1124 09:10:09.827861 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-npmfc-config-zcbpb" event={"ID":"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4","Type":"ContainerDied","Data":"de79cb55674ca0207fc7a59b27c802ecc027ac030257f5a33d496e3c2b02b09a"} Nov 24 09:10:10 crc kubenswrapper[4944]: I1124 09:10:10.839535 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"5d2fbee82caba7d0ca08ca333bcbf3e804351883721f48d8d9bfaac47a78f960"} Nov 24 09:10:10 crc kubenswrapper[4944]: I1124 09:10:10.839588 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"91b30be5efcaf3b6adec37e7c1e859fed2c520997d002c00b9500d39d521b0ed"} Nov 24 09:10:10 crc kubenswrapper[4944]: I1124 09:10:10.839605 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"e533818c9ecf15f88341998771d0639c9cac4cfb451d5374ad40bec2e815a3ff"} Nov 24 09:10:11 crc kubenswrapper[4944]: I1124 09:10:11.850008 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"c17e6f2fe0c81b82e5511368c8268b7225365160cc97761f7dc0ee157f7a8484"} Nov 24 09:10:12 crc kubenswrapper[4944]: I1124 09:10:12.338158 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-npmfc" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.308394 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.488002 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run-ovn\") pod \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.488074 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run\") pod \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.488141 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-additional-scripts\") pod \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.488185 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpzvg\" (UniqueName: \"kubernetes.io/projected/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-kube-api-access-tpzvg\") pod \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.488254 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-log-ovn\") pod \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.488251 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4" (UID: "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.488369 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-scripts\") pod \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\" (UID: \"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4\") " Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.488401 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4" (UID: "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.488776 4944 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.488802 4944 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.489093 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4" (UID: "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.489158 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run" (OuterVolumeSpecName: "var-run") pod "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4" (UID: "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.489469 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-scripts" (OuterVolumeSpecName: "scripts") pod "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4" (UID: "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.494278 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-kube-api-access-tpzvg" (OuterVolumeSpecName: "kube-api-access-tpzvg") pod "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4" (UID: "302a4ebf-b0a0-408c-8eb9-93009e2d9ec4"). InnerVolumeSpecName "kube-api-access-tpzvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.590230 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.590262 4944 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.590272 4944 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.590281 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpzvg\" (UniqueName: \"kubernetes.io/projected/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4-kube-api-access-tpzvg\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.874248 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-npmfc-config-zcbpb" event={"ID":"302a4ebf-b0a0-408c-8eb9-93009e2d9ec4","Type":"ContainerDied","Data":"78f08b95c177bc88f756e1a35fe2247319880a774db5ee06e28e464c5c5d83c2"} Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.874282 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc-config-zcbpb" Nov 24 09:10:14 crc kubenswrapper[4944]: I1124 09:10:14.874291 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78f08b95c177bc88f756e1a35fe2247319880a774db5ee06e28e464c5c5d83c2" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.422100 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-npmfc-config-zcbpb"] Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.437668 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-npmfc-config-zcbpb"] Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.551414 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-npmfc-config-w9g8d"] Nov 24 09:10:15 crc kubenswrapper[4944]: E1124 09:10:15.552557 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="302a4ebf-b0a0-408c-8eb9-93009e2d9ec4" containerName="ovn-config" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.552588 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="302a4ebf-b0a0-408c-8eb9-93009e2d9ec4" containerName="ovn-config" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.553161 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="302a4ebf-b0a0-408c-8eb9-93009e2d9ec4" containerName="ovn-config" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.554819 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.562012 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-npmfc-config-w9g8d"] Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.563323 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.728525 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.728572 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-additional-scripts\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.728609 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-scripts\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.728627 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-log-ovn\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.728849 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run-ovn\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.729223 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf6rc\" (UniqueName: \"kubernetes.io/projected/7636f393-7dc4-457e-98da-20c2ad7fc8d9-kube-api-access-tf6rc\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.831151 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run-ovn\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.831238 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf6rc\" (UniqueName: \"kubernetes.io/projected/7636f393-7dc4-457e-98da-20c2ad7fc8d9-kube-api-access-tf6rc\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.831292 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-additional-scripts\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.831308 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.831340 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-scripts\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.831362 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-log-ovn\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.831767 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-log-ovn\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.831883 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run-ovn\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.831884 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.833004 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-additional-scripts\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.833608 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-scripts\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.851827 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf6rc\" (UniqueName: \"kubernetes.io/projected/7636f393-7dc4-457e-98da-20c2ad7fc8d9-kube-api-access-tf6rc\") pod \"ovn-controller-npmfc-config-w9g8d\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:15 crc kubenswrapper[4944]: I1124 09:10:15.880511 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.292287 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="302a4ebf-b0a0-408c-8eb9-93009e2d9ec4" path="/var/lib/kubelet/pods/302a4ebf-b0a0-408c-8eb9-93009e2d9ec4/volumes" Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.382174 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-npmfc-config-w9g8d"] Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.904663 4944 generic.go:334] "Generic (PLEG): container finished" podID="7636f393-7dc4-457e-98da-20c2ad7fc8d9" containerID="6d6ec38d97812dfd2d8aba09a7804fe182ba21ae83c5a722630f8851be56f865" exitCode=0 Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.904832 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-npmfc-config-w9g8d" event={"ID":"7636f393-7dc4-457e-98da-20c2ad7fc8d9","Type":"ContainerDied","Data":"6d6ec38d97812dfd2d8aba09a7804fe182ba21ae83c5a722630f8851be56f865"} Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.905000 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-npmfc-config-w9g8d" event={"ID":"7636f393-7dc4-457e-98da-20c2ad7fc8d9","Type":"ContainerStarted","Data":"0a9fdc39edcd3cce36b13085b1392b11527468f7c379d64052ff7b4099247f9e"} Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.914040 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"48d0fed6b1c1123b8f3c416f72272d6749327411fdfbf987f7ee43dd423e071e"} Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.914107 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"6ccdc9b46a378806550b2c6500f4fd45536785660f1ba53f9c389a60f6a9bf41"} Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.914121 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"15e66884ade4e39577226708f15ca7622732f6b82d81fb79dff82cc2578f9a9a"} Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.914131 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"e6c6e71ef8c3c8177a39c99dabb96e105d42c681788bccbca30c6f9504cba707"} Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.914141 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"bb4e495759495e1efa8e3c7bf1b539d36a09c8a14b177fd703baec61783412a9"} Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.918459 4944 generic.go:334] "Generic (PLEG): container finished" podID="b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93" containerID="5f7f01bba3bbc83c216d9707bc817704d3d12a1909dfb8fada43a600b9229191" exitCode=0 Nov 24 09:10:16 crc kubenswrapper[4944]: I1124 09:10:16.918477 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vlctx" event={"ID":"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93","Type":"ContainerDied","Data":"5f7f01bba3bbc83c216d9707bc817704d3d12a1909dfb8fada43a600b9229191"} Nov 24 09:10:17 crc kubenswrapper[4944]: I1124 09:10:17.930836 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"d6d5476d01856b9678b39efa998de2c30d0cc6772d172d0bf1aa500eb8ebaca7"} Nov 24 09:10:17 crc kubenswrapper[4944]: I1124 09:10:17.930889 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerStarted","Data":"be411b5a8274b541c609635739f60faee6efa6abf3fbf8c48eed07bef8944385"} Nov 24 09:10:17 crc kubenswrapper[4944]: I1124 09:10:17.967240 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=23.463618221 podStartE2EDuration="33.967214997s" podCreationTimestamp="2025-11-24 09:09:44 +0000 UTC" firstStartedPulling="2025-11-24 09:10:05.317336286 +0000 UTC m=+1065.851776748" lastFinishedPulling="2025-11-24 09:10:15.820933062 +0000 UTC m=+1076.355373524" observedRunningTime="2025-11-24 09:10:17.964511191 +0000 UTC m=+1078.498951673" watchObservedRunningTime="2025-11-24 09:10:17.967214997 +0000 UTC m=+1078.501655469" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.303725 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.353711 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-c6zc8"] Nov 24 09:10:18 crc kubenswrapper[4944]: E1124 09:10:18.355833 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7636f393-7dc4-457e-98da-20c2ad7fc8d9" containerName="ovn-config" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.355859 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7636f393-7dc4-457e-98da-20c2ad7fc8d9" containerName="ovn-config" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.356083 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="7636f393-7dc4-457e-98da-20c2ad7fc8d9" containerName="ovn-config" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.356943 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.365485 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.372696 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-c6zc8"] Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.384732 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vlctx" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.483478 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-scripts\") pod \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.483592 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-log-ovn\") pod \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.483662 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7636f393-7dc4-457e-98da-20c2ad7fc8d9" (UID: "7636f393-7dc4-457e-98da-20c2ad7fc8d9"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.483684 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-config-data\") pod \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.483753 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zf98\" (UniqueName: \"kubernetes.io/projected/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-kube-api-access-7zf98\") pod \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.483814 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run\") pod \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.483894 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf6rc\" (UniqueName: \"kubernetes.io/projected/7636f393-7dc4-457e-98da-20c2ad7fc8d9-kube-api-access-tf6rc\") pod \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.483943 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-additional-scripts\") pod \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.483976 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-db-sync-config-data\") pod \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.484037 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-combined-ca-bundle\") pod \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\" (UID: \"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93\") " Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.484106 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run-ovn\") pod \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\" (UID: \"7636f393-7dc4-457e-98da-20c2ad7fc8d9\") " Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.484331 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.484360 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.484410 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.484461 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-config\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.484531 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gtd6\" (UniqueName: \"kubernetes.io/projected/2a750139-b28d-45e9-9cb3-0159595bf95a-kube-api-access-2gtd6\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.484605 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.484679 4944 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.485080 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "7636f393-7dc4-457e-98da-20c2ad7fc8d9" (UID: "7636f393-7dc4-457e-98da-20c2ad7fc8d9"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.485557 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-scripts" (OuterVolumeSpecName: "scripts") pod "7636f393-7dc4-457e-98da-20c2ad7fc8d9" (UID: "7636f393-7dc4-457e-98da-20c2ad7fc8d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.485967 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7636f393-7dc4-457e-98da-20c2ad7fc8d9" (UID: "7636f393-7dc4-457e-98da-20c2ad7fc8d9"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.486074 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run" (OuterVolumeSpecName: "var-run") pod "7636f393-7dc4-457e-98da-20c2ad7fc8d9" (UID: "7636f393-7dc4-457e-98da-20c2ad7fc8d9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.490061 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93" (UID: "b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.490703 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-kube-api-access-7zf98" (OuterVolumeSpecName: "kube-api-access-7zf98") pod "b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93" (UID: "b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93"). InnerVolumeSpecName "kube-api-access-7zf98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.490916 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7636f393-7dc4-457e-98da-20c2ad7fc8d9-kube-api-access-tf6rc" (OuterVolumeSpecName: "kube-api-access-tf6rc") pod "7636f393-7dc4-457e-98da-20c2ad7fc8d9" (UID: "7636f393-7dc4-457e-98da-20c2ad7fc8d9"). InnerVolumeSpecName "kube-api-access-tf6rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.512928 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93" (UID: "b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.533613 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-config-data" (OuterVolumeSpecName: "config-data") pod "b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93" (UID: "b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.586511 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-config\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587344 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-config\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587465 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gtd6\" (UniqueName: \"kubernetes.io/projected/2a750139-b28d-45e9-9cb3-0159595bf95a-kube-api-access-2gtd6\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587533 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587566 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587584 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587615 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587678 4944 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587692 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587706 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587717 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zf98\" (UniqueName: \"kubernetes.io/projected/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-kube-api-access-7zf98\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587727 4944 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7636f393-7dc4-457e-98da-20c2ad7fc8d9-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587736 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf6rc\" (UniqueName: \"kubernetes.io/projected/7636f393-7dc4-457e-98da-20c2ad7fc8d9-kube-api-access-tf6rc\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587745 4944 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7636f393-7dc4-457e-98da-20c2ad7fc8d9-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587753 4944 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.587762 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.588299 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.589148 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.590275 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.590617 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.608659 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gtd6\" (UniqueName: \"kubernetes.io/projected/2a750139-b28d-45e9-9cb3-0159595bf95a-kube-api-access-2gtd6\") pod \"dnsmasq-dns-5c79d794d7-c6zc8\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.699590 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.943630 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-vlctx" event={"ID":"b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93","Type":"ContainerDied","Data":"d56c29e4dbfc201dcde60401bdd060f3a7016b7ea7bbb29b97fdb2ef12644118"} Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.944187 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d56c29e4dbfc201dcde60401bdd060f3a7016b7ea7bbb29b97fdb2ef12644118" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.943673 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-vlctx" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.946070 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-npmfc-config-w9g8d" event={"ID":"7636f393-7dc4-457e-98da-20c2ad7fc8d9","Type":"ContainerDied","Data":"0a9fdc39edcd3cce36b13085b1392b11527468f7c379d64052ff7b4099247f9e"} Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.946089 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc-config-w9g8d" Nov 24 09:10:18 crc kubenswrapper[4944]: I1124 09:10:18.946144 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a9fdc39edcd3cce36b13085b1392b11527468f7c379d64052ff7b4099247f9e" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.134978 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-c6zc8"] Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.153267 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.407830 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-npmfc-config-w9g8d"] Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.416545 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-npmfc-config-w9g8d"] Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.444252 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-c6zc8"] Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.472435 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-4b5pb"] Nov 24 09:10:19 crc kubenswrapper[4944]: E1124 09:10:19.472778 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93" containerName="glance-db-sync" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.472796 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93" containerName="glance-db-sync" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.472963 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93" containerName="glance-db-sync" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.473836 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.497128 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-4b5pb"] Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.643995 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk4lz\" (UniqueName: \"kubernetes.io/projected/fc2988bc-f594-4018-b585-67800dc75b9b-kube-api-access-sk4lz\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.644107 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.644159 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.644186 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-config\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.644217 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.644259 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.729207 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.745846 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk4lz\" (UniqueName: \"kubernetes.io/projected/fc2988bc-f594-4018-b585-67800dc75b9b-kube-api-access-sk4lz\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.745982 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.746081 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.746120 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-config\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.746176 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.746238 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.747481 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.747627 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.747649 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.747868 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.748399 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-config\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.775191 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk4lz\" (UniqueName: \"kubernetes.io/projected/fc2988bc-f594-4018-b585-67800dc75b9b-kube-api-access-sk4lz\") pod \"dnsmasq-dns-5f59b8f679-4b5pb\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.788080 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.968107 4944 generic.go:334] "Generic (PLEG): container finished" podID="2a750139-b28d-45e9-9cb3-0159595bf95a" containerID="7f56f1a253c2473a6aa995cf6223551ec08f172d572dc44f42f26db19a07df62" exitCode=0 Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.968548 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" event={"ID":"2a750139-b28d-45e9-9cb3-0159595bf95a","Type":"ContainerDied","Data":"7f56f1a253c2473a6aa995cf6223551ec08f172d572dc44f42f26db19a07df62"} Nov 24 09:10:19 crc kubenswrapper[4944]: I1124 09:10:19.968588 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" event={"ID":"2a750139-b28d-45e9-9cb3-0159595bf95a","Type":"ContainerStarted","Data":"6c0598732693b3671ecd3b462885760568a1695b74b84ecce03dfecf5528e1ba"} Nov 24 09:10:20 crc kubenswrapper[4944]: E1124 09:10:20.167454 4944 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 24 09:10:20 crc kubenswrapper[4944]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/2a750139-b28d-45e9-9cb3-0159595bf95a/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 24 09:10:20 crc kubenswrapper[4944]: > podSandboxID="6c0598732693b3671ecd3b462885760568a1695b74b84ecce03dfecf5528e1ba" Nov 24 09:10:20 crc kubenswrapper[4944]: E1124 09:10:20.167877 4944 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 24 09:10:20 crc kubenswrapper[4944]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n97h57bh654h659h5b6hbfhc4h689h565h578h56ch8dh8bh67fhf7h5f8hc7h5d4h5d5h5f7h687h5cbh5c5h5d8h68fh669h588h59bh5c6h674h5c8h5d7q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2gtd6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5c79d794d7-c6zc8_openstack(2a750139-b28d-45e9-9cb3-0159595bf95a): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/2a750139-b28d-45e9-9cb3-0159595bf95a/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 24 09:10:20 crc kubenswrapper[4944]: > logger="UnhandledError" Nov 24 09:10:20 crc kubenswrapper[4944]: E1124 09:10:20.169029 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/2a750139-b28d-45e9-9cb3-0159595bf95a/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" podUID="2a750139-b28d-45e9-9cb3-0159595bf95a" Nov 24 09:10:20 crc kubenswrapper[4944]: I1124 09:10:20.268650 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-4b5pb"] Nov 24 09:10:20 crc kubenswrapper[4944]: I1124 09:10:20.291421 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7636f393-7dc4-457e-98da-20c2ad7fc8d9" path="/var/lib/kubelet/pods/7636f393-7dc4-457e-98da-20c2ad7fc8d9/volumes" Nov 24 09:10:20 crc kubenswrapper[4944]: I1124 09:10:20.979339 4944 generic.go:334] "Generic (PLEG): container finished" podID="fc2988bc-f594-4018-b585-67800dc75b9b" containerID="650775d8fb82494dc6f1f67ef4f93ee0286d62e5144131ad3ffafef35bd52569" exitCode=0 Nov 24 09:10:20 crc kubenswrapper[4944]: I1124 09:10:20.980512 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" event={"ID":"fc2988bc-f594-4018-b585-67800dc75b9b","Type":"ContainerDied","Data":"650775d8fb82494dc6f1f67ef4f93ee0286d62e5144131ad3ffafef35bd52569"} Nov 24 09:10:20 crc kubenswrapper[4944]: I1124 09:10:20.980551 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" event={"ID":"fc2988bc-f594-4018-b585-67800dc75b9b","Type":"ContainerStarted","Data":"99b48245f2018fca11fde69aa3674bd6276a03636b114739e54a8c0bc8c2d3fb"} Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.291686 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-b8fcn"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.295579 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b8fcn" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.303480 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-b8fcn"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.332785 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-575d-account-create-c8gbm"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.334151 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-575d-account-create-c8gbm" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.340697 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.357726 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-575d-account-create-c8gbm"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.371855 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94hgr\" (UniqueName: \"kubernetes.io/projected/a3235d75-d12c-4656-adb4-d0d485f2b45b-kube-api-access-94hgr\") pod \"cinder-db-create-b8fcn\" (UID: \"a3235d75-d12c-4656-adb4-d0d485f2b45b\") " pod="openstack/cinder-db-create-b8fcn" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.371922 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3235d75-d12c-4656-adb4-d0d485f2b45b-operator-scripts\") pod \"cinder-db-create-b8fcn\" (UID: \"a3235d75-d12c-4656-adb4-d0d485f2b45b\") " pod="openstack/cinder-db-create-b8fcn" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.379135 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-xjbgk"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.382366 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xjbgk" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.386467 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xjbgk"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.462009 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.480232 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94hgr\" (UniqueName: \"kubernetes.io/projected/a3235d75-d12c-4656-adb4-d0d485f2b45b-kube-api-access-94hgr\") pod \"cinder-db-create-b8fcn\" (UID: \"a3235d75-d12c-4656-adb4-d0d485f2b45b\") " pod="openstack/cinder-db-create-b8fcn" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.480285 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f8b9684-d40c-4c61-aff8-59008e970331-operator-scripts\") pod \"barbican-db-create-xjbgk\" (UID: \"0f8b9684-d40c-4c61-aff8-59008e970331\") " pod="openstack/barbican-db-create-xjbgk" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.480322 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3235d75-d12c-4656-adb4-d0d485f2b45b-operator-scripts\") pod \"cinder-db-create-b8fcn\" (UID: \"a3235d75-d12c-4656-adb4-d0d485f2b45b\") " pod="openstack/cinder-db-create-b8fcn" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.480358 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7851f98-142f-4123-be3b-c811c294954e-operator-scripts\") pod \"cinder-575d-account-create-c8gbm\" (UID: \"e7851f98-142f-4123-be3b-c811c294954e\") " pod="openstack/cinder-575d-account-create-c8gbm" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.480417 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmv2s\" (UniqueName: \"kubernetes.io/projected/e7851f98-142f-4123-be3b-c811c294954e-kube-api-access-vmv2s\") pod \"cinder-575d-account-create-c8gbm\" (UID: \"e7851f98-142f-4123-be3b-c811c294954e\") " pod="openstack/cinder-575d-account-create-c8gbm" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.480472 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgwcv\" (UniqueName: \"kubernetes.io/projected/0f8b9684-d40c-4c61-aff8-59008e970331-kube-api-access-wgwcv\") pod \"barbican-db-create-xjbgk\" (UID: \"0f8b9684-d40c-4c61-aff8-59008e970331\") " pod="openstack/barbican-db-create-xjbgk" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.480555 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-d270-account-create-trwxh"] Nov 24 09:10:21 crc kubenswrapper[4944]: E1124 09:10:21.480925 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a750139-b28d-45e9-9cb3-0159595bf95a" containerName="init" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.480945 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a750139-b28d-45e9-9cb3-0159595bf95a" containerName="init" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.481207 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a750139-b28d-45e9-9cb3-0159595bf95a" containerName="init" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.489610 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3235d75-d12c-4656-adb4-d0d485f2b45b-operator-scripts\") pod \"cinder-db-create-b8fcn\" (UID: \"a3235d75-d12c-4656-adb4-d0d485f2b45b\") " pod="openstack/cinder-db-create-b8fcn" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.490709 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d270-account-create-trwxh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.493551 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.494840 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d270-account-create-trwxh"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.528796 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94hgr\" (UniqueName: \"kubernetes.io/projected/a3235d75-d12c-4656-adb4-d0d485f2b45b-kube-api-access-94hgr\") pod \"cinder-db-create-b8fcn\" (UID: \"a3235d75-d12c-4656-adb4-d0d485f2b45b\") " pod="openstack/cinder-db-create-b8fcn" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.540189 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-75jn8"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.541399 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.545361 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.545803 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.545949 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qztjd" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.546106 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.566703 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-75jn8"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.581486 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-nb\") pod \"2a750139-b28d-45e9-9cb3-0159595bf95a\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.581590 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-svc\") pod \"2a750139-b28d-45e9-9cb3-0159595bf95a\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.581656 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-swift-storage-0\") pod \"2a750139-b28d-45e9-9cb3-0159595bf95a\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.581698 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gtd6\" (UniqueName: \"kubernetes.io/projected/2a750139-b28d-45e9-9cb3-0159595bf95a-kube-api-access-2gtd6\") pod \"2a750139-b28d-45e9-9cb3-0159595bf95a\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.581775 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-sb\") pod \"2a750139-b28d-45e9-9cb3-0159595bf95a\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.581840 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-config\") pod \"2a750139-b28d-45e9-9cb3-0159595bf95a\" (UID: \"2a750139-b28d-45e9-9cb3-0159595bf95a\") " Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.582095 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f8b9684-d40c-4c61-aff8-59008e970331-operator-scripts\") pod \"barbican-db-create-xjbgk\" (UID: \"0f8b9684-d40c-4c61-aff8-59008e970331\") " pod="openstack/barbican-db-create-xjbgk" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.582168 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7851f98-142f-4123-be3b-c811c294954e-operator-scripts\") pod \"cinder-575d-account-create-c8gbm\" (UID: \"e7851f98-142f-4123-be3b-c811c294954e\") " pod="openstack/cinder-575d-account-create-c8gbm" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.582227 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmv2s\" (UniqueName: \"kubernetes.io/projected/e7851f98-142f-4123-be3b-c811c294954e-kube-api-access-vmv2s\") pod \"cinder-575d-account-create-c8gbm\" (UID: \"e7851f98-142f-4123-be3b-c811c294954e\") " pod="openstack/cinder-575d-account-create-c8gbm" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.582300 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgwcv\" (UniqueName: \"kubernetes.io/projected/0f8b9684-d40c-4c61-aff8-59008e970331-kube-api-access-wgwcv\") pod \"barbican-db-create-xjbgk\" (UID: \"0f8b9684-d40c-4c61-aff8-59008e970331\") " pod="openstack/barbican-db-create-xjbgk" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.582350 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f08ed10-9454-4781-8a6b-000100a436dc-operator-scripts\") pod \"barbican-d270-account-create-trwxh\" (UID: \"5f08ed10-9454-4781-8a6b-000100a436dc\") " pod="openstack/barbican-d270-account-create-trwxh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.582392 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5bsk\" (UniqueName: \"kubernetes.io/projected/5f08ed10-9454-4781-8a6b-000100a436dc-kube-api-access-g5bsk\") pod \"barbican-d270-account-create-trwxh\" (UID: \"5f08ed10-9454-4781-8a6b-000100a436dc\") " pod="openstack/barbican-d270-account-create-trwxh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.584977 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-97djh"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.586104 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-97djh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.586484 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7851f98-142f-4123-be3b-c811c294954e-operator-scripts\") pod \"cinder-575d-account-create-c8gbm\" (UID: \"e7851f98-142f-4123-be3b-c811c294954e\") " pod="openstack/cinder-575d-account-create-c8gbm" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.590269 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f8b9684-d40c-4c61-aff8-59008e970331-operator-scripts\") pod \"barbican-db-create-xjbgk\" (UID: \"0f8b9684-d40c-4c61-aff8-59008e970331\") " pod="openstack/barbican-db-create-xjbgk" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.600342 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-97djh"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.611474 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmv2s\" (UniqueName: \"kubernetes.io/projected/e7851f98-142f-4123-be3b-c811c294954e-kube-api-access-vmv2s\") pod \"cinder-575d-account-create-c8gbm\" (UID: \"e7851f98-142f-4123-be3b-c811c294954e\") " pod="openstack/cinder-575d-account-create-c8gbm" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.613750 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a750139-b28d-45e9-9cb3-0159595bf95a-kube-api-access-2gtd6" (OuterVolumeSpecName: "kube-api-access-2gtd6") pod "2a750139-b28d-45e9-9cb3-0159595bf95a" (UID: "2a750139-b28d-45e9-9cb3-0159595bf95a"). InnerVolumeSpecName "kube-api-access-2gtd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.628988 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgwcv\" (UniqueName: \"kubernetes.io/projected/0f8b9684-d40c-4c61-aff8-59008e970331-kube-api-access-wgwcv\") pod \"barbican-db-create-xjbgk\" (UID: \"0f8b9684-d40c-4c61-aff8-59008e970331\") " pod="openstack/barbican-db-create-xjbgk" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.631414 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b8fcn" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.671193 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-575d-account-create-c8gbm" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.686266 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-combined-ca-bundle\") pod \"keystone-db-sync-75jn8\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.686388 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-config" (OuterVolumeSpecName: "config") pod "2a750139-b28d-45e9-9cb3-0159595bf95a" (UID: "2a750139-b28d-45e9-9cb3-0159595bf95a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.690644 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2a750139-b28d-45e9-9cb3-0159595bf95a" (UID: "2a750139-b28d-45e9-9cb3-0159595bf95a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.690718 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9vts\" (UniqueName: \"kubernetes.io/projected/19055ab7-c910-494d-9a31-49aadf127c69-kube-api-access-v9vts\") pod \"keystone-db-sync-75jn8\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.690920 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-operator-scripts\") pod \"neutron-db-create-97djh\" (UID: \"8dc3bfd5-b619-4b27-842a-5c5d47375c6a\") " pod="openstack/neutron-db-create-97djh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.690973 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tllgs\" (UniqueName: \"kubernetes.io/projected/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-kube-api-access-tllgs\") pod \"neutron-db-create-97djh\" (UID: \"8dc3bfd5-b619-4b27-842a-5c5d47375c6a\") " pod="openstack/neutron-db-create-97djh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.691001 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f08ed10-9454-4781-8a6b-000100a436dc-operator-scripts\") pod \"barbican-d270-account-create-trwxh\" (UID: \"5f08ed10-9454-4781-8a6b-000100a436dc\") " pod="openstack/barbican-d270-account-create-trwxh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.691058 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-config-data\") pod \"keystone-db-sync-75jn8\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.691089 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5bsk\" (UniqueName: \"kubernetes.io/projected/5f08ed10-9454-4781-8a6b-000100a436dc-kube-api-access-g5bsk\") pod \"barbican-d270-account-create-trwxh\" (UID: \"5f08ed10-9454-4781-8a6b-000100a436dc\") " pod="openstack/barbican-d270-account-create-trwxh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.691187 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gtd6\" (UniqueName: \"kubernetes.io/projected/2a750139-b28d-45e9-9cb3-0159595bf95a-kube-api-access-2gtd6\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.691215 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.691227 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.691651 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f08ed10-9454-4781-8a6b-000100a436dc-operator-scripts\") pod \"barbican-d270-account-create-trwxh\" (UID: \"5f08ed10-9454-4781-8a6b-000100a436dc\") " pod="openstack/barbican-d270-account-create-trwxh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.703992 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2a750139-b28d-45e9-9cb3-0159595bf95a" (UID: "2a750139-b28d-45e9-9cb3-0159595bf95a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.704686 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2a750139-b28d-45e9-9cb3-0159595bf95a" (UID: "2a750139-b28d-45e9-9cb3-0159595bf95a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.707654 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2a750139-b28d-45e9-9cb3-0159595bf95a" (UID: "2a750139-b28d-45e9-9cb3-0159595bf95a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.710337 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7ac6-account-create-t66tz"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.712123 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7ac6-account-create-t66tz" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.727488 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.727999 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5bsk\" (UniqueName: \"kubernetes.io/projected/5f08ed10-9454-4781-8a6b-000100a436dc-kube-api-access-g5bsk\") pod \"barbican-d270-account-create-trwxh\" (UID: \"5f08ed10-9454-4781-8a6b-000100a436dc\") " pod="openstack/barbican-d270-account-create-trwxh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.754788 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xjbgk" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.794017 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d176dce-ca46-4a21-801e-42055ae16d1d-operator-scripts\") pod \"neutron-7ac6-account-create-t66tz\" (UID: \"4d176dce-ca46-4a21-801e-42055ae16d1d\") " pod="openstack/neutron-7ac6-account-create-t66tz" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.794224 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-operator-scripts\") pod \"neutron-db-create-97djh\" (UID: \"8dc3bfd5-b619-4b27-842a-5c5d47375c6a\") " pod="openstack/neutron-db-create-97djh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.794261 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tllgs\" (UniqueName: \"kubernetes.io/projected/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-kube-api-access-tllgs\") pod \"neutron-db-create-97djh\" (UID: \"8dc3bfd5-b619-4b27-842a-5c5d47375c6a\") " pod="openstack/neutron-db-create-97djh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.794288 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-config-data\") pod \"keystone-db-sync-75jn8\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.794317 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-combined-ca-bundle\") pod \"keystone-db-sync-75jn8\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.794333 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w792\" (UniqueName: \"kubernetes.io/projected/4d176dce-ca46-4a21-801e-42055ae16d1d-kube-api-access-9w792\") pod \"neutron-7ac6-account-create-t66tz\" (UID: \"4d176dce-ca46-4a21-801e-42055ae16d1d\") " pod="openstack/neutron-7ac6-account-create-t66tz" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.794366 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9vts\" (UniqueName: \"kubernetes.io/projected/19055ab7-c910-494d-9a31-49aadf127c69-kube-api-access-v9vts\") pod \"keystone-db-sync-75jn8\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.794406 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.794418 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.794426 4944 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a750139-b28d-45e9-9cb3-0159595bf95a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.795432 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-operator-scripts\") pod \"neutron-db-create-97djh\" (UID: \"8dc3bfd5-b619-4b27-842a-5c5d47375c6a\") " pod="openstack/neutron-db-create-97djh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.815117 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7ac6-account-create-t66tz"] Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.846858 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-config-data\") pod \"keystone-db-sync-75jn8\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.847346 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-combined-ca-bundle\") pod \"keystone-db-sync-75jn8\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.847718 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tllgs\" (UniqueName: \"kubernetes.io/projected/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-kube-api-access-tllgs\") pod \"neutron-db-create-97djh\" (UID: \"8dc3bfd5-b619-4b27-842a-5c5d47375c6a\") " pod="openstack/neutron-db-create-97djh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.855941 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d270-account-create-trwxh" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.900955 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w792\" (UniqueName: \"kubernetes.io/projected/4d176dce-ca46-4a21-801e-42055ae16d1d-kube-api-access-9w792\") pod \"neutron-7ac6-account-create-t66tz\" (UID: \"4d176dce-ca46-4a21-801e-42055ae16d1d\") " pod="openstack/neutron-7ac6-account-create-t66tz" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.901082 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d176dce-ca46-4a21-801e-42055ae16d1d-operator-scripts\") pod \"neutron-7ac6-account-create-t66tz\" (UID: \"4d176dce-ca46-4a21-801e-42055ae16d1d\") " pod="openstack/neutron-7ac6-account-create-t66tz" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.902013 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d176dce-ca46-4a21-801e-42055ae16d1d-operator-scripts\") pod \"neutron-7ac6-account-create-t66tz\" (UID: \"4d176dce-ca46-4a21-801e-42055ae16d1d\") " pod="openstack/neutron-7ac6-account-create-t66tz" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.902542 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9vts\" (UniqueName: \"kubernetes.io/projected/19055ab7-c910-494d-9a31-49aadf127c69-kube-api-access-v9vts\") pod \"keystone-db-sync-75jn8\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:21 crc kubenswrapper[4944]: I1124 09:10:21.929709 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w792\" (UniqueName: \"kubernetes.io/projected/4d176dce-ca46-4a21-801e-42055ae16d1d-kube-api-access-9w792\") pod \"neutron-7ac6-account-create-t66tz\" (UID: \"4d176dce-ca46-4a21-801e-42055ae16d1d\") " pod="openstack/neutron-7ac6-account-create-t66tz" Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.005261 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.028650 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-97djh" Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.029555 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.030613 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-c6zc8" event={"ID":"2a750139-b28d-45e9-9cb3-0159595bf95a","Type":"ContainerDied","Data":"6c0598732693b3671ecd3b462885760568a1695b74b84ecce03dfecf5528e1ba"} Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.030676 4944 scope.go:117] "RemoveContainer" containerID="7f56f1a253c2473a6aa995cf6223551ec08f172d572dc44f42f26db19a07df62" Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.034229 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" event={"ID":"fc2988bc-f594-4018-b585-67800dc75b9b","Type":"ContainerStarted","Data":"dad14e323f57ede01cb229731af038ba2f3cae867b2341ccd19c185119305a67"} Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.037135 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.050299 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7ac6-account-create-t66tz" Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.104516 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" podStartSLOduration=3.104442507 podStartE2EDuration="3.104442507s" podCreationTimestamp="2025-11-24 09:10:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:10:22.075441813 +0000 UTC m=+1082.609882275" watchObservedRunningTime="2025-11-24 09:10:22.104442507 +0000 UTC m=+1082.638882969" Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.174946 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-c6zc8"] Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.185156 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-c6zc8"] Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.298734 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a750139-b28d-45e9-9cb3-0159595bf95a" path="/var/lib/kubelet/pods/2a750139-b28d-45e9-9cb3-0159595bf95a/volumes" Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.409413 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-575d-account-create-c8gbm"] Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.415436 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-b8fcn"] Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.505975 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xjbgk"] Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.687492 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d270-account-create-trwxh"] Nov 24 09:10:22 crc kubenswrapper[4944]: W1124 09:10:22.697731 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f08ed10_9454_4781_8a6b_000100a436dc.slice/crio-6568170b4d6bb822d992cdc0e443b8767898fe153304b0093f83ff933d7b7bee WatchSource:0}: Error finding container 6568170b4d6bb822d992cdc0e443b8767898fe153304b0093f83ff933d7b7bee: Status 404 returned error can't find the container with id 6568170b4d6bb822d992cdc0e443b8767898fe153304b0093f83ff933d7b7bee Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.790917 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7ac6-account-create-t66tz"] Nov 24 09:10:22 crc kubenswrapper[4944]: W1124 09:10:22.802246 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d176dce_ca46_4a21_801e_42055ae16d1d.slice/crio-b7deec761be75fa876f9dd2905e89a0e2f4861b5fbb0cf9662d03ead99bbe595 WatchSource:0}: Error finding container b7deec761be75fa876f9dd2905e89a0e2f4861b5fbb0cf9662d03ead99bbe595: Status 404 returned error can't find the container with id b7deec761be75fa876f9dd2905e89a0e2f4861b5fbb0cf9662d03ead99bbe595 Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.804116 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-97djh"] Nov 24 09:10:22 crc kubenswrapper[4944]: I1124 09:10:22.814780 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-75jn8"] Nov 24 09:10:22 crc kubenswrapper[4944]: W1124 09:10:22.831118 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19055ab7_c910_494d_9a31_49aadf127c69.slice/crio-b1ac55daa193cf809f089c96e9f293075a02d483a822fa6bb349f5a4f23f0c8e WatchSource:0}: Error finding container b1ac55daa193cf809f089c96e9f293075a02d483a822fa6bb349f5a4f23f0c8e: Status 404 returned error can't find the container with id b1ac55daa193cf809f089c96e9f293075a02d483a822fa6bb349f5a4f23f0c8e Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.044343 4944 generic.go:334] "Generic (PLEG): container finished" podID="e7851f98-142f-4123-be3b-c811c294954e" containerID="879f66b45dd1486ce94200c6d646a5c12a6b2404828999d8fa88dd82bb7f9229" exitCode=0 Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.044444 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-575d-account-create-c8gbm" event={"ID":"e7851f98-142f-4123-be3b-c811c294954e","Type":"ContainerDied","Data":"879f66b45dd1486ce94200c6d646a5c12a6b2404828999d8fa88dd82bb7f9229"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.044476 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-575d-account-create-c8gbm" event={"ID":"e7851f98-142f-4123-be3b-c811c294954e","Type":"ContainerStarted","Data":"f7ee35e5c0e6cbf1c052b842ad56b7d9d1e24ee86a259bd92ea0b5291b83a8ee"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.046277 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d270-account-create-trwxh" event={"ID":"5f08ed10-9454-4781-8a6b-000100a436dc","Type":"ContainerStarted","Data":"1bfe5877d95a636388b958f1f7d0d9a81c2b7c523ca095ed7bbdf4c983a6a8b6"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.046304 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d270-account-create-trwxh" event={"ID":"5f08ed10-9454-4781-8a6b-000100a436dc","Type":"ContainerStarted","Data":"6568170b4d6bb822d992cdc0e443b8767898fe153304b0093f83ff933d7b7bee"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.053736 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-97djh" event={"ID":"8dc3bfd5-b619-4b27-842a-5c5d47375c6a","Type":"ContainerStarted","Data":"05f3827d7409a75ca27686ed4e2f7544dca88296220b6c81cd7eceee4395c8cb"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.053782 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-97djh" event={"ID":"8dc3bfd5-b619-4b27-842a-5c5d47375c6a","Type":"ContainerStarted","Data":"ac61aaeabb6e202e631ec2cebeb51d35a1ed8f59d46ec316169e7ef2beb366d1"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.064534 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7ac6-account-create-t66tz" event={"ID":"4d176dce-ca46-4a21-801e-42055ae16d1d","Type":"ContainerStarted","Data":"0f0f741121c6d302e3545b02a6d4a19fa67ea57a811079fb4057d58abb0f5764"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.064578 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7ac6-account-create-t66tz" event={"ID":"4d176dce-ca46-4a21-801e-42055ae16d1d","Type":"ContainerStarted","Data":"b7deec761be75fa876f9dd2905e89a0e2f4861b5fbb0cf9662d03ead99bbe595"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.071267 4944 generic.go:334] "Generic (PLEG): container finished" podID="a3235d75-d12c-4656-adb4-d0d485f2b45b" containerID="e3ebe45315d452b608c18183ab8a4f20e94a8a7e1e85df1008d16d4287b10042" exitCode=0 Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.071325 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b8fcn" event={"ID":"a3235d75-d12c-4656-adb4-d0d485f2b45b","Type":"ContainerDied","Data":"e3ebe45315d452b608c18183ab8a4f20e94a8a7e1e85df1008d16d4287b10042"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.071347 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b8fcn" event={"ID":"a3235d75-d12c-4656-adb4-d0d485f2b45b","Type":"ContainerStarted","Data":"23ec05258fe05a089d299c8edbfdbd731bf290d0247f6086f85ec2c60d16810a"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.078175 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75jn8" event={"ID":"19055ab7-c910-494d-9a31-49aadf127c69","Type":"ContainerStarted","Data":"b1ac55daa193cf809f089c96e9f293075a02d483a822fa6bb349f5a4f23f0c8e"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.088733 4944 generic.go:334] "Generic (PLEG): container finished" podID="0f8b9684-d40c-4c61-aff8-59008e970331" containerID="59887e3dd78e9b25b49b5075672aa68784184f607941ee0b247b3554533b3f0e" exitCode=0 Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.088825 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xjbgk" event={"ID":"0f8b9684-d40c-4c61-aff8-59008e970331","Type":"ContainerDied","Data":"59887e3dd78e9b25b49b5075672aa68784184f607941ee0b247b3554533b3f0e"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.088874 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xjbgk" event={"ID":"0f8b9684-d40c-4c61-aff8-59008e970331","Type":"ContainerStarted","Data":"c899176c5acde187085288a03f5238ee33ebde57a886953eaf426fd389848710"} Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.091705 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-97djh" podStartSLOduration=2.091690572 podStartE2EDuration="2.091690572s" podCreationTimestamp="2025-11-24 09:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:10:23.08252406 +0000 UTC m=+1083.616964522" watchObservedRunningTime="2025-11-24 09:10:23.091690572 +0000 UTC m=+1083.626131034" Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.120244 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-d270-account-create-trwxh" podStartSLOduration=2.120192961 podStartE2EDuration="2.120192961s" podCreationTimestamp="2025-11-24 09:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:10:23.117943589 +0000 UTC m=+1083.652384051" watchObservedRunningTime="2025-11-24 09:10:23.120192961 +0000 UTC m=+1083.654633423" Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.120514 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7ac6-account-create-t66tz" podStartSLOduration=2.120510311 podStartE2EDuration="2.120510311s" podCreationTimestamp="2025-11-24 09:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:10:23.101164294 +0000 UTC m=+1083.635604866" watchObservedRunningTime="2025-11-24 09:10:23.120510311 +0000 UTC m=+1083.654950773" Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.549062 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:10:23 crc kubenswrapper[4944]: I1124 09:10:23.549409 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.105908 4944 generic.go:334] "Generic (PLEG): container finished" podID="4d176dce-ca46-4a21-801e-42055ae16d1d" containerID="0f0f741121c6d302e3545b02a6d4a19fa67ea57a811079fb4057d58abb0f5764" exitCode=0 Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.106003 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7ac6-account-create-t66tz" event={"ID":"4d176dce-ca46-4a21-801e-42055ae16d1d","Type":"ContainerDied","Data":"0f0f741121c6d302e3545b02a6d4a19fa67ea57a811079fb4057d58abb0f5764"} Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.107509 4944 generic.go:334] "Generic (PLEG): container finished" podID="5f08ed10-9454-4781-8a6b-000100a436dc" containerID="1bfe5877d95a636388b958f1f7d0d9a81c2b7c523ca095ed7bbdf4c983a6a8b6" exitCode=0 Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.107542 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d270-account-create-trwxh" event={"ID":"5f08ed10-9454-4781-8a6b-000100a436dc","Type":"ContainerDied","Data":"1bfe5877d95a636388b958f1f7d0d9a81c2b7c523ca095ed7bbdf4c983a6a8b6"} Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.109175 4944 generic.go:334] "Generic (PLEG): container finished" podID="8dc3bfd5-b619-4b27-842a-5c5d47375c6a" containerID="05f3827d7409a75ca27686ed4e2f7544dca88296220b6c81cd7eceee4395c8cb" exitCode=0 Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.109366 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-97djh" event={"ID":"8dc3bfd5-b619-4b27-842a-5c5d47375c6a","Type":"ContainerDied","Data":"05f3827d7409a75ca27686ed4e2f7544dca88296220b6c81cd7eceee4395c8cb"} Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.610202 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-575d-account-create-c8gbm" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.617153 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xjbgk" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.624019 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b8fcn" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.663347 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgwcv\" (UniqueName: \"kubernetes.io/projected/0f8b9684-d40c-4c61-aff8-59008e970331-kube-api-access-wgwcv\") pod \"0f8b9684-d40c-4c61-aff8-59008e970331\" (UID: \"0f8b9684-d40c-4c61-aff8-59008e970331\") " Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.663411 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7851f98-142f-4123-be3b-c811c294954e-operator-scripts\") pod \"e7851f98-142f-4123-be3b-c811c294954e\" (UID: \"e7851f98-142f-4123-be3b-c811c294954e\") " Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.663615 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f8b9684-d40c-4c61-aff8-59008e970331-operator-scripts\") pod \"0f8b9684-d40c-4c61-aff8-59008e970331\" (UID: \"0f8b9684-d40c-4c61-aff8-59008e970331\") " Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.663680 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmv2s\" (UniqueName: \"kubernetes.io/projected/e7851f98-142f-4123-be3b-c811c294954e-kube-api-access-vmv2s\") pod \"e7851f98-142f-4123-be3b-c811c294954e\" (UID: \"e7851f98-142f-4123-be3b-c811c294954e\") " Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.667581 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f8b9684-d40c-4c61-aff8-59008e970331-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f8b9684-d40c-4c61-aff8-59008e970331" (UID: "0f8b9684-d40c-4c61-aff8-59008e970331"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.667871 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7851f98-142f-4123-be3b-c811c294954e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e7851f98-142f-4123-be3b-c811c294954e" (UID: "e7851f98-142f-4123-be3b-c811c294954e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.700626 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7851f98-142f-4123-be3b-c811c294954e-kube-api-access-vmv2s" (OuterVolumeSpecName: "kube-api-access-vmv2s") pod "e7851f98-142f-4123-be3b-c811c294954e" (UID: "e7851f98-142f-4123-be3b-c811c294954e"). InnerVolumeSpecName "kube-api-access-vmv2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.701828 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f8b9684-d40c-4c61-aff8-59008e970331-kube-api-access-wgwcv" (OuterVolumeSpecName: "kube-api-access-wgwcv") pod "0f8b9684-d40c-4c61-aff8-59008e970331" (UID: "0f8b9684-d40c-4c61-aff8-59008e970331"). InnerVolumeSpecName "kube-api-access-wgwcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.764679 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3235d75-d12c-4656-adb4-d0d485f2b45b-operator-scripts\") pod \"a3235d75-d12c-4656-adb4-d0d485f2b45b\" (UID: \"a3235d75-d12c-4656-adb4-d0d485f2b45b\") " Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.764738 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94hgr\" (UniqueName: \"kubernetes.io/projected/a3235d75-d12c-4656-adb4-d0d485f2b45b-kube-api-access-94hgr\") pod \"a3235d75-d12c-4656-adb4-d0d485f2b45b\" (UID: \"a3235d75-d12c-4656-adb4-d0d485f2b45b\") " Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.765313 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f8b9684-d40c-4c61-aff8-59008e970331-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.765333 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmv2s\" (UniqueName: \"kubernetes.io/projected/e7851f98-142f-4123-be3b-c811c294954e-kube-api-access-vmv2s\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.765344 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgwcv\" (UniqueName: \"kubernetes.io/projected/0f8b9684-d40c-4c61-aff8-59008e970331-kube-api-access-wgwcv\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.765354 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7851f98-142f-4123-be3b-c811c294954e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.765697 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3235d75-d12c-4656-adb4-d0d485f2b45b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3235d75-d12c-4656-adb4-d0d485f2b45b" (UID: "a3235d75-d12c-4656-adb4-d0d485f2b45b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.768460 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3235d75-d12c-4656-adb4-d0d485f2b45b-kube-api-access-94hgr" (OuterVolumeSpecName: "kube-api-access-94hgr") pod "a3235d75-d12c-4656-adb4-d0d485f2b45b" (UID: "a3235d75-d12c-4656-adb4-d0d485f2b45b"). InnerVolumeSpecName "kube-api-access-94hgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.866503 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3235d75-d12c-4656-adb4-d0d485f2b45b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:24 crc kubenswrapper[4944]: I1124 09:10:24.866536 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94hgr\" (UniqueName: \"kubernetes.io/projected/a3235d75-d12c-4656-adb4-d0d485f2b45b-kube-api-access-94hgr\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:25 crc kubenswrapper[4944]: I1124 09:10:25.119427 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b8fcn" Nov 24 09:10:25 crc kubenswrapper[4944]: I1124 09:10:25.119421 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b8fcn" event={"ID":"a3235d75-d12c-4656-adb4-d0d485f2b45b","Type":"ContainerDied","Data":"23ec05258fe05a089d299c8edbfdbd731bf290d0247f6086f85ec2c60d16810a"} Nov 24 09:10:25 crc kubenswrapper[4944]: I1124 09:10:25.119929 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23ec05258fe05a089d299c8edbfdbd731bf290d0247f6086f85ec2c60d16810a" Nov 24 09:10:25 crc kubenswrapper[4944]: I1124 09:10:25.121021 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xjbgk" Nov 24 09:10:25 crc kubenswrapper[4944]: I1124 09:10:25.121032 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xjbgk" event={"ID":"0f8b9684-d40c-4c61-aff8-59008e970331","Type":"ContainerDied","Data":"c899176c5acde187085288a03f5238ee33ebde57a886953eaf426fd389848710"} Nov 24 09:10:25 crc kubenswrapper[4944]: I1124 09:10:25.121092 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c899176c5acde187085288a03f5238ee33ebde57a886953eaf426fd389848710" Nov 24 09:10:25 crc kubenswrapper[4944]: I1124 09:10:25.122487 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-575d-account-create-c8gbm" event={"ID":"e7851f98-142f-4123-be3b-c811c294954e","Type":"ContainerDied","Data":"f7ee35e5c0e6cbf1c052b842ad56b7d9d1e24ee86a259bd92ea0b5291b83a8ee"} Nov 24 09:10:25 crc kubenswrapper[4944]: I1124 09:10:25.122524 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7ee35e5c0e6cbf1c052b842ad56b7d9d1e24ee86a259bd92ea0b5291b83a8ee" Nov 24 09:10:25 crc kubenswrapper[4944]: I1124 09:10:25.122628 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-575d-account-create-c8gbm" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.603628 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d270-account-create-trwxh" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.611995 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-97djh" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.620183 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7ac6-account-create-t66tz" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.722560 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d176dce-ca46-4a21-801e-42055ae16d1d-operator-scripts\") pod \"4d176dce-ca46-4a21-801e-42055ae16d1d\" (UID: \"4d176dce-ca46-4a21-801e-42055ae16d1d\") " Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.722635 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-operator-scripts\") pod \"8dc3bfd5-b619-4b27-842a-5c5d47375c6a\" (UID: \"8dc3bfd5-b619-4b27-842a-5c5d47375c6a\") " Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.722742 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f08ed10-9454-4781-8a6b-000100a436dc-operator-scripts\") pod \"5f08ed10-9454-4781-8a6b-000100a436dc\" (UID: \"5f08ed10-9454-4781-8a6b-000100a436dc\") " Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.722840 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tllgs\" (UniqueName: \"kubernetes.io/projected/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-kube-api-access-tllgs\") pod \"8dc3bfd5-b619-4b27-842a-5c5d47375c6a\" (UID: \"8dc3bfd5-b619-4b27-842a-5c5d47375c6a\") " Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.722872 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5bsk\" (UniqueName: \"kubernetes.io/projected/5f08ed10-9454-4781-8a6b-000100a436dc-kube-api-access-g5bsk\") pod \"5f08ed10-9454-4781-8a6b-000100a436dc\" (UID: \"5f08ed10-9454-4781-8a6b-000100a436dc\") " Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.722985 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w792\" (UniqueName: \"kubernetes.io/projected/4d176dce-ca46-4a21-801e-42055ae16d1d-kube-api-access-9w792\") pod \"4d176dce-ca46-4a21-801e-42055ae16d1d\" (UID: \"4d176dce-ca46-4a21-801e-42055ae16d1d\") " Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.723810 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f08ed10-9454-4781-8a6b-000100a436dc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5f08ed10-9454-4781-8a6b-000100a436dc" (UID: "5f08ed10-9454-4781-8a6b-000100a436dc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.724238 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d176dce-ca46-4a21-801e-42055ae16d1d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d176dce-ca46-4a21-801e-42055ae16d1d" (UID: "4d176dce-ca46-4a21-801e-42055ae16d1d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.724234 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8dc3bfd5-b619-4b27-842a-5c5d47375c6a" (UID: "8dc3bfd5-b619-4b27-842a-5c5d47375c6a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.730384 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f08ed10-9454-4781-8a6b-000100a436dc-kube-api-access-g5bsk" (OuterVolumeSpecName: "kube-api-access-g5bsk") pod "5f08ed10-9454-4781-8a6b-000100a436dc" (UID: "5f08ed10-9454-4781-8a6b-000100a436dc"). InnerVolumeSpecName "kube-api-access-g5bsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.732813 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-kube-api-access-tllgs" (OuterVolumeSpecName: "kube-api-access-tllgs") pod "8dc3bfd5-b619-4b27-842a-5c5d47375c6a" (UID: "8dc3bfd5-b619-4b27-842a-5c5d47375c6a"). InnerVolumeSpecName "kube-api-access-tllgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.738219 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d176dce-ca46-4a21-801e-42055ae16d1d-kube-api-access-9w792" (OuterVolumeSpecName: "kube-api-access-9w792") pod "4d176dce-ca46-4a21-801e-42055ae16d1d" (UID: "4d176dce-ca46-4a21-801e-42055ae16d1d"). InnerVolumeSpecName "kube-api-access-9w792". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.825560 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w792\" (UniqueName: \"kubernetes.io/projected/4d176dce-ca46-4a21-801e-42055ae16d1d-kube-api-access-9w792\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.825594 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d176dce-ca46-4a21-801e-42055ae16d1d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.825604 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.825612 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f08ed10-9454-4781-8a6b-000100a436dc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.825621 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tllgs\" (UniqueName: \"kubernetes.io/projected/8dc3bfd5-b619-4b27-842a-5c5d47375c6a-kube-api-access-tllgs\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:27 crc kubenswrapper[4944]: I1124 09:10:27.825629 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5bsk\" (UniqueName: \"kubernetes.io/projected/5f08ed10-9454-4781-8a6b-000100a436dc-kube-api-access-g5bsk\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:28 crc kubenswrapper[4944]: I1124 09:10:28.151856 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-97djh" event={"ID":"8dc3bfd5-b619-4b27-842a-5c5d47375c6a","Type":"ContainerDied","Data":"ac61aaeabb6e202e631ec2cebeb51d35a1ed8f59d46ec316169e7ef2beb366d1"} Nov 24 09:10:28 crc kubenswrapper[4944]: I1124 09:10:28.151904 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac61aaeabb6e202e631ec2cebeb51d35a1ed8f59d46ec316169e7ef2beb366d1" Nov 24 09:10:28 crc kubenswrapper[4944]: I1124 09:10:28.151931 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-97djh" Nov 24 09:10:28 crc kubenswrapper[4944]: I1124 09:10:28.154428 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7ac6-account-create-t66tz" event={"ID":"4d176dce-ca46-4a21-801e-42055ae16d1d","Type":"ContainerDied","Data":"b7deec761be75fa876f9dd2905e89a0e2f4861b5fbb0cf9662d03ead99bbe595"} Nov 24 09:10:28 crc kubenswrapper[4944]: I1124 09:10:28.154481 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7ac6-account-create-t66tz" Nov 24 09:10:28 crc kubenswrapper[4944]: I1124 09:10:28.154485 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7deec761be75fa876f9dd2905e89a0e2f4861b5fbb0cf9662d03ead99bbe595" Nov 24 09:10:28 crc kubenswrapper[4944]: I1124 09:10:28.156024 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d270-account-create-trwxh" event={"ID":"5f08ed10-9454-4781-8a6b-000100a436dc","Type":"ContainerDied","Data":"6568170b4d6bb822d992cdc0e443b8767898fe153304b0093f83ff933d7b7bee"} Nov 24 09:10:28 crc kubenswrapper[4944]: I1124 09:10:28.156072 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6568170b4d6bb822d992cdc0e443b8767898fe153304b0093f83ff933d7b7bee" Nov 24 09:10:28 crc kubenswrapper[4944]: I1124 09:10:28.156107 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d270-account-create-trwxh" Nov 24 09:10:29 crc kubenswrapper[4944]: I1124 09:10:29.167433 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75jn8" event={"ID":"19055ab7-c910-494d-9a31-49aadf127c69","Type":"ContainerStarted","Data":"e734c391470a420ba500ddc3044001dc8a942a79f4c8290fe6f87c79d611a796"} Nov 24 09:10:29 crc kubenswrapper[4944]: I1124 09:10:29.188484 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-75jn8" podStartSLOduration=2.536898372 podStartE2EDuration="8.188462335s" podCreationTimestamp="2025-11-24 09:10:21 +0000 UTC" firstStartedPulling="2025-11-24 09:10:22.83966824 +0000 UTC m=+1083.374108702" lastFinishedPulling="2025-11-24 09:10:28.491232203 +0000 UTC m=+1089.025672665" observedRunningTime="2025-11-24 09:10:29.18077507 +0000 UTC m=+1089.715215542" watchObservedRunningTime="2025-11-24 09:10:29.188462335 +0000 UTC m=+1089.722902797" Nov 24 09:10:29 crc kubenswrapper[4944]: I1124 09:10:29.790659 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:29 crc kubenswrapper[4944]: I1124 09:10:29.855852 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9ct72"] Nov 24 09:10:29 crc kubenswrapper[4944]: I1124 09:10:29.856489 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" podUID="1330f372-b8dc-455e-9ef5-ddeb01f6df00" containerName="dnsmasq-dns" containerID="cri-o://c54d394785155844676a6c1d096e0bb324d50ac7e9818cd6cbe354818c5132b5" gracePeriod=10 Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.182012 4944 generic.go:334] "Generic (PLEG): container finished" podID="1330f372-b8dc-455e-9ef5-ddeb01f6df00" containerID="c54d394785155844676a6c1d096e0bb324d50ac7e9818cd6cbe354818c5132b5" exitCode=0 Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.182094 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" event={"ID":"1330f372-b8dc-455e-9ef5-ddeb01f6df00","Type":"ContainerDied","Data":"c54d394785155844676a6c1d096e0bb324d50ac7e9818cd6cbe354818c5132b5"} Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.313351 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.472352 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh9kq\" (UniqueName: \"kubernetes.io/projected/1330f372-b8dc-455e-9ef5-ddeb01f6df00-kube-api-access-nh9kq\") pod \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.472484 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-config\") pod \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.472564 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-dns-svc\") pod \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.472596 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-nb\") pod \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.472644 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-sb\") pod \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\" (UID: \"1330f372-b8dc-455e-9ef5-ddeb01f6df00\") " Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.490247 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1330f372-b8dc-455e-9ef5-ddeb01f6df00-kube-api-access-nh9kq" (OuterVolumeSpecName: "kube-api-access-nh9kq") pod "1330f372-b8dc-455e-9ef5-ddeb01f6df00" (UID: "1330f372-b8dc-455e-9ef5-ddeb01f6df00"). InnerVolumeSpecName "kube-api-access-nh9kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.521361 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-config" (OuterVolumeSpecName: "config") pod "1330f372-b8dc-455e-9ef5-ddeb01f6df00" (UID: "1330f372-b8dc-455e-9ef5-ddeb01f6df00"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.522591 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1330f372-b8dc-455e-9ef5-ddeb01f6df00" (UID: "1330f372-b8dc-455e-9ef5-ddeb01f6df00"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.523715 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1330f372-b8dc-455e-9ef5-ddeb01f6df00" (UID: "1330f372-b8dc-455e-9ef5-ddeb01f6df00"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.525974 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1330f372-b8dc-455e-9ef5-ddeb01f6df00" (UID: "1330f372-b8dc-455e-9ef5-ddeb01f6df00"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.573974 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.574013 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.574029 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.574063 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1330f372-b8dc-455e-9ef5-ddeb01f6df00-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:30 crc kubenswrapper[4944]: I1124 09:10:30.574079 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh9kq\" (UniqueName: \"kubernetes.io/projected/1330f372-b8dc-455e-9ef5-ddeb01f6df00-kube-api-access-nh9kq\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:31 crc kubenswrapper[4944]: I1124 09:10:31.191679 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" event={"ID":"1330f372-b8dc-455e-9ef5-ddeb01f6df00","Type":"ContainerDied","Data":"3d1320536f688fa86f8708aa1fbb55d9632f44fe22431ef74dedb3d03369572c"} Nov 24 09:10:31 crc kubenswrapper[4944]: I1124 09:10:31.191732 4944 scope.go:117] "RemoveContainer" containerID="c54d394785155844676a6c1d096e0bb324d50ac7e9818cd6cbe354818c5132b5" Nov 24 09:10:31 crc kubenswrapper[4944]: I1124 09:10:31.191877 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-9ct72" Nov 24 09:10:31 crc kubenswrapper[4944]: I1124 09:10:31.217695 4944 scope.go:117] "RemoveContainer" containerID="67563f065abdc1b0605307ff47b7a0c04feebb212462f56b94c8960913c9012e" Nov 24 09:10:31 crc kubenswrapper[4944]: I1124 09:10:31.226898 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9ct72"] Nov 24 09:10:31 crc kubenswrapper[4944]: I1124 09:10:31.232785 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9ct72"] Nov 24 09:10:32 crc kubenswrapper[4944]: I1124 09:10:32.286643 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1330f372-b8dc-455e-9ef5-ddeb01f6df00" path="/var/lib/kubelet/pods/1330f372-b8dc-455e-9ef5-ddeb01f6df00/volumes" Nov 24 09:10:33 crc kubenswrapper[4944]: I1124 09:10:33.208913 4944 generic.go:334] "Generic (PLEG): container finished" podID="19055ab7-c910-494d-9a31-49aadf127c69" containerID="e734c391470a420ba500ddc3044001dc8a942a79f4c8290fe6f87c79d611a796" exitCode=0 Nov 24 09:10:33 crc kubenswrapper[4944]: I1124 09:10:33.208954 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75jn8" event={"ID":"19055ab7-c910-494d-9a31-49aadf127c69","Type":"ContainerDied","Data":"e734c391470a420ba500ddc3044001dc8a942a79f4c8290fe6f87c79d611a796"} Nov 24 09:10:34 crc kubenswrapper[4944]: I1124 09:10:34.518316 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:34 crc kubenswrapper[4944]: I1124 09:10:34.639207 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-combined-ca-bundle\") pod \"19055ab7-c910-494d-9a31-49aadf127c69\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " Nov 24 09:10:34 crc kubenswrapper[4944]: I1124 09:10:34.639733 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-config-data\") pod \"19055ab7-c910-494d-9a31-49aadf127c69\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " Nov 24 09:10:34 crc kubenswrapper[4944]: I1124 09:10:34.639781 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9vts\" (UniqueName: \"kubernetes.io/projected/19055ab7-c910-494d-9a31-49aadf127c69-kube-api-access-v9vts\") pod \"19055ab7-c910-494d-9a31-49aadf127c69\" (UID: \"19055ab7-c910-494d-9a31-49aadf127c69\") " Nov 24 09:10:34 crc kubenswrapper[4944]: I1124 09:10:34.644709 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19055ab7-c910-494d-9a31-49aadf127c69-kube-api-access-v9vts" (OuterVolumeSpecName: "kube-api-access-v9vts") pod "19055ab7-c910-494d-9a31-49aadf127c69" (UID: "19055ab7-c910-494d-9a31-49aadf127c69"). InnerVolumeSpecName "kube-api-access-v9vts". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:34 crc kubenswrapper[4944]: I1124 09:10:34.666532 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19055ab7-c910-494d-9a31-49aadf127c69" (UID: "19055ab7-c910-494d-9a31-49aadf127c69"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:34 crc kubenswrapper[4944]: I1124 09:10:34.685180 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-config-data" (OuterVolumeSpecName: "config-data") pod "19055ab7-c910-494d-9a31-49aadf127c69" (UID: "19055ab7-c910-494d-9a31-49aadf127c69"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:34 crc kubenswrapper[4944]: I1124 09:10:34.741610 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:34 crc kubenswrapper[4944]: I1124 09:10:34.741680 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9vts\" (UniqueName: \"kubernetes.io/projected/19055ab7-c910-494d-9a31-49aadf127c69-kube-api-access-v9vts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:34 crc kubenswrapper[4944]: I1124 09:10:34.741695 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19055ab7-c910-494d-9a31-49aadf127c69-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.225086 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75jn8" event={"ID":"19055ab7-c910-494d-9a31-49aadf127c69","Type":"ContainerDied","Data":"b1ac55daa193cf809f089c96e9f293075a02d483a822fa6bb349f5a4f23f0c8e"} Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.225153 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1ac55daa193cf809f089c96e9f293075a02d483a822fa6bb349f5a4f23f0c8e" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.225166 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75jn8" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.489414 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-5b65k"] Nov 24 09:10:35 crc kubenswrapper[4944]: E1124 09:10:35.490031 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1330f372-b8dc-455e-9ef5-ddeb01f6df00" containerName="dnsmasq-dns" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490078 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1330f372-b8dc-455e-9ef5-ddeb01f6df00" containerName="dnsmasq-dns" Nov 24 09:10:35 crc kubenswrapper[4944]: E1124 09:10:35.490108 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3235d75-d12c-4656-adb4-d0d485f2b45b" containerName="mariadb-database-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490117 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3235d75-d12c-4656-adb4-d0d485f2b45b" containerName="mariadb-database-create" Nov 24 09:10:35 crc kubenswrapper[4944]: E1124 09:10:35.490134 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7851f98-142f-4123-be3b-c811c294954e" containerName="mariadb-account-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490145 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7851f98-142f-4123-be3b-c811c294954e" containerName="mariadb-account-create" Nov 24 09:10:35 crc kubenswrapper[4944]: E1124 09:10:35.490163 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f8b9684-d40c-4c61-aff8-59008e970331" containerName="mariadb-database-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490173 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f8b9684-d40c-4c61-aff8-59008e970331" containerName="mariadb-database-create" Nov 24 09:10:35 crc kubenswrapper[4944]: E1124 09:10:35.490188 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1330f372-b8dc-455e-9ef5-ddeb01f6df00" containerName="init" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490198 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1330f372-b8dc-455e-9ef5-ddeb01f6df00" containerName="init" Nov 24 09:10:35 crc kubenswrapper[4944]: E1124 09:10:35.490214 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc3bfd5-b619-4b27-842a-5c5d47375c6a" containerName="mariadb-database-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490222 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc3bfd5-b619-4b27-842a-5c5d47375c6a" containerName="mariadb-database-create" Nov 24 09:10:35 crc kubenswrapper[4944]: E1124 09:10:35.490242 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f08ed10-9454-4781-8a6b-000100a436dc" containerName="mariadb-account-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490249 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f08ed10-9454-4781-8a6b-000100a436dc" containerName="mariadb-account-create" Nov 24 09:10:35 crc kubenswrapper[4944]: E1124 09:10:35.490261 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d176dce-ca46-4a21-801e-42055ae16d1d" containerName="mariadb-account-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490269 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d176dce-ca46-4a21-801e-42055ae16d1d" containerName="mariadb-account-create" Nov 24 09:10:35 crc kubenswrapper[4944]: E1124 09:10:35.490280 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19055ab7-c910-494d-9a31-49aadf127c69" containerName="keystone-db-sync" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490287 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="19055ab7-c910-494d-9a31-49aadf127c69" containerName="keystone-db-sync" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490464 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="19055ab7-c910-494d-9a31-49aadf127c69" containerName="keystone-db-sync" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490475 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f08ed10-9454-4781-8a6b-000100a436dc" containerName="mariadb-account-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490488 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1330f372-b8dc-455e-9ef5-ddeb01f6df00" containerName="dnsmasq-dns" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490508 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f8b9684-d40c-4c61-aff8-59008e970331" containerName="mariadb-database-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490521 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dc3bfd5-b619-4b27-842a-5c5d47375c6a" containerName="mariadb-database-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490532 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3235d75-d12c-4656-adb4-d0d485f2b45b" containerName="mariadb-database-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490539 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7851f98-142f-4123-be3b-c811c294954e" containerName="mariadb-account-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.490570 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d176dce-ca46-4a21-801e-42055ae16d1d" containerName="mariadb-account-create" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.491683 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.510122 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5kzx5"] Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.511305 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.516257 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.516294 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.516505 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.516515 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qztjd" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.516593 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.521615 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-5b65k"] Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555033 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-scripts\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555145 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn5hp\" (UniqueName: \"kubernetes.io/projected/a2d5a94f-767f-4203-bf57-9172eccdce70-kube-api-access-dn5hp\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555208 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555225 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-config-data\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555249 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkg48\" (UniqueName: \"kubernetes.io/projected/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-kube-api-access-bkg48\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555268 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-credential-keys\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555337 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555357 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-config\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555377 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-combined-ca-bundle\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555397 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555422 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555457 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-fernet-keys\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.555718 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5kzx5"] Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.656908 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkg48\" (UniqueName: \"kubernetes.io/projected/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-kube-api-access-bkg48\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.657255 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-credential-keys\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.657298 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.657320 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-config\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.657339 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-combined-ca-bundle\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.657360 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.657387 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.657424 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-fernet-keys\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.657461 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-scripts\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.657488 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn5hp\" (UniqueName: \"kubernetes.io/projected/a2d5a94f-767f-4203-bf57-9172eccdce70-kube-api-access-dn5hp\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.657521 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.657544 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-config-data\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.658815 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.661193 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-config\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.661782 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.662743 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.663447 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-config-data\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.665911 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.666610 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-credential-keys\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.672666 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-combined-ca-bundle\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.681805 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-fernet-keys\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.683412 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-scripts\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.714712 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkg48\" (UniqueName: \"kubernetes.io/projected/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-kube-api-access-bkg48\") pod \"keystone-bootstrap-5kzx5\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.739860 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn5hp\" (UniqueName: \"kubernetes.io/projected/a2d5a94f-767f-4203-bf57-9172eccdce70-kube-api-access-dn5hp\") pod \"dnsmasq-dns-bbf5cc879-5b65k\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.810673 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.835362 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.916200 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-nst6k"] Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.920502 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nst6k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.926745 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qtvxl" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.927635 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.937227 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-nst6k"] Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.969301 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-vqngq"] Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.970626 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.983084 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-vk9ck" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.983289 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.983410 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.987019 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-db-sync-config-data\") pod \"barbican-db-sync-nst6k\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " pod="openstack/barbican-db-sync-nst6k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.987149 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-combined-ca-bundle\") pod \"barbican-db-sync-nst6k\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " pod="openstack/barbican-db-sync-nst6k" Nov 24 09:10:35 crc kubenswrapper[4944]: I1124 09:10:35.987219 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhz6v\" (UniqueName: \"kubernetes.io/projected/0493ffa7-e203-4f40-b117-07b07fe13380-kube-api-access-mhz6v\") pod \"barbican-db-sync-nst6k\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " pod="openstack/barbican-db-sync-nst6k" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.054482 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vqngq"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.091109 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-s7k5s"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.092080 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-combined-ca-bundle\") pod \"neutron-db-sync-vqngq\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.092144 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-config\") pod \"neutron-db-sync-vqngq\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.092193 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.092199 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-combined-ca-bundle\") pod \"barbican-db-sync-nst6k\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " pod="openstack/barbican-db-sync-nst6k" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.092286 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhz6v\" (UniqueName: \"kubernetes.io/projected/0493ffa7-e203-4f40-b117-07b07fe13380-kube-api-access-mhz6v\") pod \"barbican-db-sync-nst6k\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " pod="openstack/barbican-db-sync-nst6k" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.092318 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x54mf\" (UniqueName: \"kubernetes.io/projected/587209cc-b2bb-460e-a6d4-1391b6294864-kube-api-access-x54mf\") pod \"neutron-db-sync-vqngq\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.092349 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-db-sync-config-data\") pod \"barbican-db-sync-nst6k\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " pod="openstack/barbican-db-sync-nst6k" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.096557 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9lk6q" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.096859 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.097000 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.098756 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-db-sync-config-data\") pod \"barbican-db-sync-nst6k\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " pod="openstack/barbican-db-sync-nst6k" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.115155 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-combined-ca-bundle\") pod \"barbican-db-sync-nst6k\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " pod="openstack/barbican-db-sync-nst6k" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.126761 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-s7k5s"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.139098 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-hpjlj"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.140218 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.146162 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.149004 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-j5rrf" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.149190 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.151864 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hpjlj"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.168469 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-5b65k"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.176748 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhz6v\" (UniqueName: \"kubernetes.io/projected/0493ffa7-e203-4f40-b117-07b07fe13380-kube-api-access-mhz6v\") pod \"barbican-db-sync-nst6k\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " pod="openstack/barbican-db-sync-nst6k" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.193994 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x54mf\" (UniqueName: \"kubernetes.io/projected/587209cc-b2bb-460e-a6d4-1391b6294864-kube-api-access-x54mf\") pod \"neutron-db-sync-vqngq\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194065 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-scripts\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194098 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c338ca6-2795-4c94-b863-23579147b584-etc-machine-id\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194117 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt4k7\" (UniqueName: \"kubernetes.io/projected/6c338ca6-2795-4c94-b863-23579147b584-kube-api-access-tt4k7\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194139 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-combined-ca-bundle\") pod \"neutron-db-sync-vqngq\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194166 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-scripts\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194352 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-db-sync-config-data\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194372 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-config\") pod \"neutron-db-sync-vqngq\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194388 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29dcx\" (UniqueName: \"kubernetes.io/projected/6b3396e1-991a-4a6e-a7a6-665658af7e63-kube-api-access-29dcx\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194404 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b3396e1-991a-4a6e-a7a6-665658af7e63-logs\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194432 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-combined-ca-bundle\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194499 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-combined-ca-bundle\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194515 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-config-data\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.194540 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-config-data\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.199858 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-combined-ca-bundle\") pod \"neutron-db-sync-vqngq\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.208118 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-config\") pod \"neutron-db-sync-vqngq\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.222242 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x54mf\" (UniqueName: \"kubernetes.io/projected/587209cc-b2bb-460e-a6d4-1391b6294864-kube-api-access-x54mf\") pod \"neutron-db-sync-vqngq\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.223006 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-62h9v"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.235666 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.237015 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-62h9v"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.258379 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.263995 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nst6k" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.282509 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.288529 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.291904 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.307324 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.307382 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-config\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.307432 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f6j9\" (UniqueName: \"kubernetes.io/projected/28e6918e-c11e-423e-996b-c43ee889155c-kube-api-access-9f6j9\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.307525 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-combined-ca-bundle\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.307544 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.307566 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-config-data\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.307620 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.307739 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-config-data\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.307903 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-scripts\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.308467 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c338ca6-2795-4c94-b863-23579147b584-etc-machine-id\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.308528 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt4k7\" (UniqueName: \"kubernetes.io/projected/6c338ca6-2795-4c94-b863-23579147b584-kube-api-access-tt4k7\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.308599 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-scripts\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.308630 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-db-sync-config-data\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.308653 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.308836 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29dcx\" (UniqueName: \"kubernetes.io/projected/6b3396e1-991a-4a6e-a7a6-665658af7e63-kube-api-access-29dcx\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.308874 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b3396e1-991a-4a6e-a7a6-665658af7e63-logs\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.308906 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-combined-ca-bundle\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.339953 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.340147 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-db-sync-config-data\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.341122 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b3396e1-991a-4a6e-a7a6-665658af7e63-logs\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.341344 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c338ca6-2795-4c94-b863-23579147b584-etc-machine-id\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.343644 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-scripts\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.345444 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-scripts\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.347957 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-combined-ca-bundle\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.348840 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-combined-ca-bundle\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.350493 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-config-data\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.350758 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-config-data\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.362128 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29dcx\" (UniqueName: \"kubernetes.io/projected/6b3396e1-991a-4a6e-a7a6-665658af7e63-kube-api-access-29dcx\") pod \"placement-db-sync-s7k5s\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.373622 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt4k7\" (UniqueName: \"kubernetes.io/projected/6c338ca6-2795-4c94-b863-23579147b584-kube-api-access-tt4k7\") pod \"cinder-db-sync-hpjlj\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.392518 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.410812 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.410866 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.410905 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.410967 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-log-httpd\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.411018 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-run-httpd\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.411033 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wt9v\" (UniqueName: \"kubernetes.io/projected/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-kube-api-access-4wt9v\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.411087 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-scripts\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.411111 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.411133 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.411189 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.411210 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-config-data\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.411227 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-config\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.411242 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f6j9\" (UniqueName: \"kubernetes.io/projected/28e6918e-c11e-423e-996b-c43ee889155c-kube-api-access-9f6j9\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.415178 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.417515 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-config\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.417577 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.419601 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.420188 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.427131 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s7k5s" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.448691 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f6j9\" (UniqueName: \"kubernetes.io/projected/28e6918e-c11e-423e-996b-c43ee889155c-kube-api-access-9f6j9\") pod \"dnsmasq-dns-56df8fb6b7-62h9v\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.508436 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.512474 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-config-data\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.512545 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.512594 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-log-httpd\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.512641 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-run-httpd\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.512665 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wt9v\" (UniqueName: \"kubernetes.io/projected/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-kube-api-access-4wt9v\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.512718 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-scripts\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.513151 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.513221 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-log-httpd\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.513719 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-run-httpd\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.517535 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-config-data\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.517607 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.528790 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.535034 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-scripts\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.537733 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wt9v\" (UniqueName: \"kubernetes.io/projected/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-kube-api-access-4wt9v\") pod \"ceilometer-0\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.565917 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.625836 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-5b65k"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.663223 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.713213 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5kzx5"] Nov 24 09:10:36 crc kubenswrapper[4944]: W1124 09:10:36.724706 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5af61f51_5798_4a0c_83a0_6dfd6d20d96a.slice/crio-9198376c5b0093e8b68c93c3ed6875c97da0d5f331ffdfcc9ce31cbb92f9a2d3 WatchSource:0}: Error finding container 9198376c5b0093e8b68c93c3ed6875c97da0d5f331ffdfcc9ce31cbb92f9a2d3: Status 404 returned error can't find the container with id 9198376c5b0093e8b68c93c3ed6875c97da0d5f331ffdfcc9ce31cbb92f9a2d3 Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.765106 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.766659 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.777239 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.777640 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-9rvqc" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.777869 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.778353 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.786556 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.819918 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.820634 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-scripts\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.820674 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.820994 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-logs\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.821038 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-config-data\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.821107 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.821352 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlfnr\" (UniqueName: \"kubernetes.io/projected/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-kube-api-access-tlfnr\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.821455 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.827710 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-nst6k"] Nov 24 09:10:36 crc kubenswrapper[4944]: W1124 09:10:36.853528 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0493ffa7_e203_4f40_b117_07b07fe13380.slice/crio-32c7d96b84faa5da891c5f38296c1230a00c462a301e85957b3c5e9ad07259f0 WatchSource:0}: Error finding container 32c7d96b84faa5da891c5f38296c1230a00c462a301e85957b3c5e9ad07259f0: Status 404 returned error can't find the container with id 32c7d96b84faa5da891c5f38296c1230a00c462a301e85957b3c5e9ad07259f0 Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.908138 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.922906 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.923001 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-logs\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.923063 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-config-data\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.923096 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.923143 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlfnr\" (UniqueName: \"kubernetes.io/projected/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-kube-api-access-tlfnr\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.923199 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.923250 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.923293 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-scripts\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.925872 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.929475 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.931107 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-logs\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.933444 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.941497 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.942854 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.955898 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.960581 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.964150 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.967921 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-config-data\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.969362 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-scripts\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.990635 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlfnr\" (UniqueName: \"kubernetes.io/projected/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-kube-api-access-tlfnr\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:36 crc kubenswrapper[4944]: I1124 09:10:36.993337 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vqngq"] Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.024262 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.024331 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.024620 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.024682 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.024715 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.024735 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.024795 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.024820 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxtjh\" (UniqueName: \"kubernetes.io/projected/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-kube-api-access-dxtjh\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.033080 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.121318 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-s7k5s"] Nov 24 09:10:37 crc kubenswrapper[4944]: W1124 09:10:37.124972 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b3396e1_991a_4a6e_a7a6_665658af7e63.slice/crio-a6e037ef5c3b909e2f83c757deaa7e41ad0c44c422292fc6fca384dfab8f39e4 WatchSource:0}: Error finding container a6e037ef5c3b909e2f83c757deaa7e41ad0c44c422292fc6fca384dfab8f39e4: Status 404 returned error can't find the container with id a6e037ef5c3b909e2f83c757deaa7e41ad0c44c422292fc6fca384dfab8f39e4 Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.126040 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.126168 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxtjh\" (UniqueName: \"kubernetes.io/projected/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-kube-api-access-dxtjh\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.126206 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.126243 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.126299 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.126361 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.126401 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.126424 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.127596 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.128155 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.130211 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.138331 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.144307 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.147983 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxtjh\" (UniqueName: \"kubernetes.io/projected/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-kube-api-access-dxtjh\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.156132 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.156454 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.186439 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.264920 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nst6k" event={"ID":"0493ffa7-e203-4f40-b117-07b07fe13380","Type":"ContainerStarted","Data":"32c7d96b84faa5da891c5f38296c1230a00c462a301e85957b3c5e9ad07259f0"} Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.270454 4944 generic.go:334] "Generic (PLEG): container finished" podID="a2d5a94f-767f-4203-bf57-9172eccdce70" containerID="411d2f6bf470d20789c3e50a7e151d8de4b31c5517815487bdc02b125f2cc2db" exitCode=0 Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.270542 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" event={"ID":"a2d5a94f-767f-4203-bf57-9172eccdce70","Type":"ContainerDied","Data":"411d2f6bf470d20789c3e50a7e151d8de4b31c5517815487bdc02b125f2cc2db"} Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.270568 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" event={"ID":"a2d5a94f-767f-4203-bf57-9172eccdce70","Type":"ContainerStarted","Data":"4082e970f1207d795b0227231f71ea54654c0904a160488b221f8bdc89bfbe69"} Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.292202 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hpjlj"] Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.316001 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vqngq" event={"ID":"587209cc-b2bb-460e-a6d4-1391b6294864","Type":"ContainerStarted","Data":"0c0565dec9e48700e3377ada79e525e8bfcaf6d1918e31822b21484621342dca"} Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.316076 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vqngq" event={"ID":"587209cc-b2bb-460e-a6d4-1391b6294864","Type":"ContainerStarted","Data":"31e43085dd5c27d1da92b39e70a0dfd51e3a7d222fa701ed81e352d17778db0a"} Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.324241 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.345737 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.347466 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-62h9v"] Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.361036 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5kzx5" event={"ID":"5af61f51-5798-4a0c-83a0-6dfd6d20d96a","Type":"ContainerStarted","Data":"645b1d9673243c219c166fe53c247425c184e6160cec86a567ca59eeea550661"} Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.361102 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5kzx5" event={"ID":"5af61f51-5798-4a0c-83a0-6dfd6d20d96a","Type":"ContainerStarted","Data":"9198376c5b0093e8b68c93c3ed6875c97da0d5f331ffdfcc9ce31cbb92f9a2d3"} Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.372508 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-vqngq" podStartSLOduration=2.372489663 podStartE2EDuration="2.372489663s" podCreationTimestamp="2025-11-24 09:10:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:10:37.346501655 +0000 UTC m=+1097.880942127" watchObservedRunningTime="2025-11-24 09:10:37.372489663 +0000 UTC m=+1097.906930115" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.389378 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5kzx5" podStartSLOduration=2.389358181 podStartE2EDuration="2.389358181s" podCreationTimestamp="2025-11-24 09:10:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:10:37.382876604 +0000 UTC m=+1097.917317086" watchObservedRunningTime="2025-11-24 09:10:37.389358181 +0000 UTC m=+1097.923798643" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.391462 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s7k5s" event={"ID":"6b3396e1-991a-4a6e-a7a6-665658af7e63","Type":"ContainerStarted","Data":"a6e037ef5c3b909e2f83c757deaa7e41ad0c44c422292fc6fca384dfab8f39e4"} Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.472555 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:10:37 crc kubenswrapper[4944]: W1124 09:10:37.522549 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde53ccf2_83b3_4116_a8eb_d5ac212e11e7.slice/crio-aaf03c72c4a7c24d8656ba98234f060e4ff324095c8d5dae0770c4fb6e763068 WatchSource:0}: Error finding container aaf03c72c4a7c24d8656ba98234f060e4ff324095c8d5dae0770c4fb6e763068: Status 404 returned error can't find the container with id aaf03c72c4a7c24d8656ba98234f060e4ff324095c8d5dae0770c4fb6e763068 Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.923522 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.955621 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-config\") pod \"a2d5a94f-767f-4203-bf57-9172eccdce70\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.955725 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-swift-storage-0\") pod \"a2d5a94f-767f-4203-bf57-9172eccdce70\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.955797 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-sb\") pod \"a2d5a94f-767f-4203-bf57-9172eccdce70\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.955851 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-nb\") pod \"a2d5a94f-767f-4203-bf57-9172eccdce70\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.955965 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-svc\") pod \"a2d5a94f-767f-4203-bf57-9172eccdce70\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.956176 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn5hp\" (UniqueName: \"kubernetes.io/projected/a2d5a94f-767f-4203-bf57-9172eccdce70-kube-api-access-dn5hp\") pod \"a2d5a94f-767f-4203-bf57-9172eccdce70\" (UID: \"a2d5a94f-767f-4203-bf57-9172eccdce70\") " Nov 24 09:10:37 crc kubenswrapper[4944]: I1124 09:10:37.965586 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2d5a94f-767f-4203-bf57-9172eccdce70-kube-api-access-dn5hp" (OuterVolumeSpecName: "kube-api-access-dn5hp") pod "a2d5a94f-767f-4203-bf57-9172eccdce70" (UID: "a2d5a94f-767f-4203-bf57-9172eccdce70"). InnerVolumeSpecName "kube-api-access-dn5hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.006884 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a2d5a94f-767f-4203-bf57-9172eccdce70" (UID: "a2d5a94f-767f-4203-bf57-9172eccdce70"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.008750 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-config" (OuterVolumeSpecName: "config") pod "a2d5a94f-767f-4203-bf57-9172eccdce70" (UID: "a2d5a94f-767f-4203-bf57-9172eccdce70"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.016316 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a2d5a94f-767f-4203-bf57-9172eccdce70" (UID: "a2d5a94f-767f-4203-bf57-9172eccdce70"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.017203 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a2d5a94f-767f-4203-bf57-9172eccdce70" (UID: "a2d5a94f-767f-4203-bf57-9172eccdce70"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.074270 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.074316 4944 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.074327 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.074338 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.074352 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn5hp\" (UniqueName: \"kubernetes.io/projected/a2d5a94f-767f-4203-bf57-9172eccdce70-kube-api-access-dn5hp\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.080613 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a2d5a94f-767f-4203-bf57-9172eccdce70" (UID: "a2d5a94f-767f-4203-bf57-9172eccdce70"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.179068 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2d5a94f-767f-4203-bf57-9172eccdce70-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.246798 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.412023 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" event={"ID":"a2d5a94f-767f-4203-bf57-9172eccdce70","Type":"ContainerDied","Data":"4082e970f1207d795b0227231f71ea54654c0904a160488b221f8bdc89bfbe69"} Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.412203 4944 scope.go:117] "RemoveContainer" containerID="411d2f6bf470d20789c3e50a7e151d8de4b31c5517815487bdc02b125f2cc2db" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.412466 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-5b65k" Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.420777 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de53ccf2-83b3-4116-a8eb-d5ac212e11e7","Type":"ContainerStarted","Data":"aaf03c72c4a7c24d8656ba98234f060e4ff324095c8d5dae0770c4fb6e763068"} Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.424154 4944 generic.go:334] "Generic (PLEG): container finished" podID="28e6918e-c11e-423e-996b-c43ee889155c" containerID="6f04f1e4430849952d1609d61a9811d5bc67da263bac5157d0480d29fa8bd109" exitCode=0 Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.424209 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" event={"ID":"28e6918e-c11e-423e-996b-c43ee889155c","Type":"ContainerDied","Data":"6f04f1e4430849952d1609d61a9811d5bc67da263bac5157d0480d29fa8bd109"} Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.424264 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" event={"ID":"28e6918e-c11e-423e-996b-c43ee889155c","Type":"ContainerStarted","Data":"fdb2abf2a0aa8c58a0a0e177d6607a5e3a4cb820cf5d78281663a7c4ce885250"} Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.425699 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd","Type":"ContainerStarted","Data":"262452f701b1841f46f9ae9fc5c06156b54eca41f5c73a3e87476dfd93624ba9"} Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.431472 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hpjlj" event={"ID":"6c338ca6-2795-4c94-b863-23579147b584","Type":"ContainerStarted","Data":"8f5e3a19fc7d82256328b30ec36865a117cd5cf5680039e095ff5b7437e8ff6e"} Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.522894 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-5b65k"] Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.531268 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-5b65k"] Nov 24 09:10:38 crc kubenswrapper[4944]: I1124 09:10:38.989422 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:10:39 crc kubenswrapper[4944]: I1124 09:10:39.055092 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:10:39 crc kubenswrapper[4944]: I1124 09:10:39.075670 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:10:39 crc kubenswrapper[4944]: I1124 09:10:39.088517 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:10:39 crc kubenswrapper[4944]: I1124 09:10:39.473988 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" event={"ID":"28e6918e-c11e-423e-996b-c43ee889155c","Type":"ContainerStarted","Data":"4c6a05fbd86035dc80b848ecd34654a1e05f03fb73dc590f509e5bdbc7cb4bc2"} Nov 24 09:10:39 crc kubenswrapper[4944]: I1124 09:10:39.474081 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:39 crc kubenswrapper[4944]: I1124 09:10:39.478127 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a0f99e5-5705-40dd-bc57-8faf9deb0aad","Type":"ContainerStarted","Data":"59a930b0ce32c8073992fb22af39476cb95da8680827f64832667159f9ddb983"} Nov 24 09:10:39 crc kubenswrapper[4944]: I1124 09:10:39.496830 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" podStartSLOduration=3.496810528 podStartE2EDuration="3.496810528s" podCreationTimestamp="2025-11-24 09:10:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:10:39.492171501 +0000 UTC m=+1100.026611983" watchObservedRunningTime="2025-11-24 09:10:39.496810528 +0000 UTC m=+1100.031250990" Nov 24 09:10:40 crc kubenswrapper[4944]: I1124 09:10:40.332961 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2d5a94f-767f-4203-bf57-9172eccdce70" path="/var/lib/kubelet/pods/a2d5a94f-767f-4203-bf57-9172eccdce70/volumes" Nov 24 09:10:40 crc kubenswrapper[4944]: I1124 09:10:40.504760 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd","Type":"ContainerStarted","Data":"03e2a6129f9e35125553e7b1098cb7459e8a4daaad033636c72180b697f6d667"} Nov 24 09:10:40 crc kubenswrapper[4944]: I1124 09:10:40.508039 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a0f99e5-5705-40dd-bc57-8faf9deb0aad","Type":"ContainerStarted","Data":"f8955f4029dd931e250c21461cd99c9ce093999fbaf9a8b7c8d006905e56b1a9"} Nov 24 09:10:41 crc kubenswrapper[4944]: I1124 09:10:41.533657 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd","Type":"ContainerStarted","Data":"440f5c50d132372c6d902d2a2f00660b621fb19a9c9e8368dd186c043e190d94"} Nov 24 09:10:41 crc kubenswrapper[4944]: I1124 09:10:41.533742 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" containerName="glance-log" containerID="cri-o://03e2a6129f9e35125553e7b1098cb7459e8a4daaad033636c72180b697f6d667" gracePeriod=30 Nov 24 09:10:41 crc kubenswrapper[4944]: I1124 09:10:41.533821 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" containerName="glance-httpd" containerID="cri-o://440f5c50d132372c6d902d2a2f00660b621fb19a9c9e8368dd186c043e190d94" gracePeriod=30 Nov 24 09:10:41 crc kubenswrapper[4944]: I1124 09:10:41.544259 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a0f99e5-5705-40dd-bc57-8faf9deb0aad","Type":"ContainerStarted","Data":"c71680738e1c9564a89a6810520284a64b66da8b6f0a69ee9842305d14ac903c"} Nov 24 09:10:41 crc kubenswrapper[4944]: I1124 09:10:41.544409 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9a0f99e5-5705-40dd-bc57-8faf9deb0aad" containerName="glance-log" containerID="cri-o://f8955f4029dd931e250c21461cd99c9ce093999fbaf9a8b7c8d006905e56b1a9" gracePeriod=30 Nov 24 09:10:41 crc kubenswrapper[4944]: I1124 09:10:41.544512 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9a0f99e5-5705-40dd-bc57-8faf9deb0aad" containerName="glance-httpd" containerID="cri-o://c71680738e1c9564a89a6810520284a64b66da8b6f0a69ee9842305d14ac903c" gracePeriod=30 Nov 24 09:10:41 crc kubenswrapper[4944]: I1124 09:10:41.559449 4944 generic.go:334] "Generic (PLEG): container finished" podID="5af61f51-5798-4a0c-83a0-6dfd6d20d96a" containerID="645b1d9673243c219c166fe53c247425c184e6160cec86a567ca59eeea550661" exitCode=0 Nov 24 09:10:41 crc kubenswrapper[4944]: I1124 09:10:41.559532 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5kzx5" event={"ID":"5af61f51-5798-4a0c-83a0-6dfd6d20d96a","Type":"ContainerDied","Data":"645b1d9673243c219c166fe53c247425c184e6160cec86a567ca59eeea550661"} Nov 24 09:10:41 crc kubenswrapper[4944]: I1124 09:10:41.575069 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.575036165 podStartE2EDuration="6.575036165s" podCreationTimestamp="2025-11-24 09:10:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:10:41.571779581 +0000 UTC m=+1102.106220053" watchObservedRunningTime="2025-11-24 09:10:41.575036165 +0000 UTC m=+1102.109476627" Nov 24 09:10:41 crc kubenswrapper[4944]: I1124 09:10:41.671607 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.671582271 podStartE2EDuration="6.671582271s" podCreationTimestamp="2025-11-24 09:10:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:10:41.663571626 +0000 UTC m=+1102.198012098" watchObservedRunningTime="2025-11-24 09:10:41.671582271 +0000 UTC m=+1102.206022733" Nov 24 09:10:42 crc kubenswrapper[4944]: I1124 09:10:42.581990 4944 generic.go:334] "Generic (PLEG): container finished" podID="f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" containerID="440f5c50d132372c6d902d2a2f00660b621fb19a9c9e8368dd186c043e190d94" exitCode=0 Nov 24 09:10:42 crc kubenswrapper[4944]: I1124 09:10:42.582020 4944 generic.go:334] "Generic (PLEG): container finished" podID="f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" containerID="03e2a6129f9e35125553e7b1098cb7459e8a4daaad033636c72180b697f6d667" exitCode=143 Nov 24 09:10:42 crc kubenswrapper[4944]: I1124 09:10:42.582091 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd","Type":"ContainerDied","Data":"440f5c50d132372c6d902d2a2f00660b621fb19a9c9e8368dd186c043e190d94"} Nov 24 09:10:42 crc kubenswrapper[4944]: I1124 09:10:42.582118 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd","Type":"ContainerDied","Data":"03e2a6129f9e35125553e7b1098cb7459e8a4daaad033636c72180b697f6d667"} Nov 24 09:10:42 crc kubenswrapper[4944]: I1124 09:10:42.584793 4944 generic.go:334] "Generic (PLEG): container finished" podID="9a0f99e5-5705-40dd-bc57-8faf9deb0aad" containerID="c71680738e1c9564a89a6810520284a64b66da8b6f0a69ee9842305d14ac903c" exitCode=0 Nov 24 09:10:42 crc kubenswrapper[4944]: I1124 09:10:42.584828 4944 generic.go:334] "Generic (PLEG): container finished" podID="9a0f99e5-5705-40dd-bc57-8faf9deb0aad" containerID="f8955f4029dd931e250c21461cd99c9ce093999fbaf9a8b7c8d006905e56b1a9" exitCode=143 Nov 24 09:10:42 crc kubenswrapper[4944]: I1124 09:10:42.584858 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a0f99e5-5705-40dd-bc57-8faf9deb0aad","Type":"ContainerDied","Data":"c71680738e1c9564a89a6810520284a64b66da8b6f0a69ee9842305d14ac903c"} Nov 24 09:10:42 crc kubenswrapper[4944]: I1124 09:10:42.584912 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a0f99e5-5705-40dd-bc57-8faf9deb0aad","Type":"ContainerDied","Data":"f8955f4029dd931e250c21461cd99c9ce093999fbaf9a8b7c8d006905e56b1a9"} Nov 24 09:10:46 crc kubenswrapper[4944]: I1124 09:10:46.568714 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:10:46 crc kubenswrapper[4944]: I1124 09:10:46.637990 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-4b5pb"] Nov 24 09:10:46 crc kubenswrapper[4944]: I1124 09:10:46.638534 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" podUID="fc2988bc-f594-4018-b585-67800dc75b9b" containerName="dnsmasq-dns" containerID="cri-o://dad14e323f57ede01cb229731af038ba2f3cae867b2341ccd19c185119305a67" gracePeriod=10 Nov 24 09:10:47 crc kubenswrapper[4944]: I1124 09:10:47.638283 4944 generic.go:334] "Generic (PLEG): container finished" podID="fc2988bc-f594-4018-b585-67800dc75b9b" containerID="dad14e323f57ede01cb229731af038ba2f3cae867b2341ccd19c185119305a67" exitCode=0 Nov 24 09:10:47 crc kubenswrapper[4944]: I1124 09:10:47.638368 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" event={"ID":"fc2988bc-f594-4018-b585-67800dc75b9b","Type":"ContainerDied","Data":"dad14e323f57ede01cb229731af038ba2f3cae867b2341ccd19c185119305a67"} Nov 24 09:10:49 crc kubenswrapper[4944]: I1124 09:10:49.789439 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" podUID="fc2988bc-f594-4018-b585-67800dc75b9b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: connect: connection refused" Nov 24 09:10:53 crc kubenswrapper[4944]: I1124 09:10:53.548695 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:10:53 crc kubenswrapper[4944]: I1124 09:10:53.549022 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.704769 4944 generic.go:334] "Generic (PLEG): container finished" podID="587209cc-b2bb-460e-a6d4-1391b6294864" containerID="0c0565dec9e48700e3377ada79e525e8bfcaf6d1918e31822b21484621342dca" exitCode=0 Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.704816 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vqngq" event={"ID":"587209cc-b2bb-460e-a6d4-1391b6294864","Type":"ContainerDied","Data":"0c0565dec9e48700e3377ada79e525e8bfcaf6d1918e31822b21484621342dca"} Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.788517 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" podUID="fc2988bc-f594-4018-b585-67800dc75b9b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: connect: connection refused" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.833942 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.840966 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-scripts\") pod \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.841047 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-credential-keys\") pod \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.841262 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkg48\" (UniqueName: \"kubernetes.io/projected/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-kube-api-access-bkg48\") pod \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.841338 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-fernet-keys\") pod \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.841360 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-config-data\") pod \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.841386 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-combined-ca-bundle\") pod \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\" (UID: \"5af61f51-5798-4a0c-83a0-6dfd6d20d96a\") " Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.847984 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-kube-api-access-bkg48" (OuterVolumeSpecName: "kube-api-access-bkg48") pod "5af61f51-5798-4a0c-83a0-6dfd6d20d96a" (UID: "5af61f51-5798-4a0c-83a0-6dfd6d20d96a"). InnerVolumeSpecName "kube-api-access-bkg48". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.863779 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5af61f51-5798-4a0c-83a0-6dfd6d20d96a" (UID: "5af61f51-5798-4a0c-83a0-6dfd6d20d96a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.880730 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-scripts" (OuterVolumeSpecName: "scripts") pod "5af61f51-5798-4a0c-83a0-6dfd6d20d96a" (UID: "5af61f51-5798-4a0c-83a0-6dfd6d20d96a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.880837 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5af61f51-5798-4a0c-83a0-6dfd6d20d96a" (UID: "5af61f51-5798-4a0c-83a0-6dfd6d20d96a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.884458 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-config-data" (OuterVolumeSpecName: "config-data") pod "5af61f51-5798-4a0c-83a0-6dfd6d20d96a" (UID: "5af61f51-5798-4a0c-83a0-6dfd6d20d96a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.894477 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5af61f51-5798-4a0c-83a0-6dfd6d20d96a" (UID: "5af61f51-5798-4a0c-83a0-6dfd6d20d96a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.943495 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkg48\" (UniqueName: \"kubernetes.io/projected/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-kube-api-access-bkg48\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.943760 4944 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.943773 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.943785 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.943796 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:54 crc kubenswrapper[4944]: I1124 09:10:54.943806 4944 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5af61f51-5798-4a0c-83a0-6dfd6d20d96a-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:55 crc kubenswrapper[4944]: I1124 09:10:55.716502 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5kzx5" Nov 24 09:10:55 crc kubenswrapper[4944]: I1124 09:10:55.718159 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5kzx5" event={"ID":"5af61f51-5798-4a0c-83a0-6dfd6d20d96a","Type":"ContainerDied","Data":"9198376c5b0093e8b68c93c3ed6875c97da0d5f331ffdfcc9ce31cbb92f9a2d3"} Nov 24 09:10:55 crc kubenswrapper[4944]: I1124 09:10:55.718201 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9198376c5b0093e8b68c93c3ed6875c97da0d5f331ffdfcc9ce31cbb92f9a2d3" Nov 24 09:10:55 crc kubenswrapper[4944]: I1124 09:10:55.927383 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5kzx5"] Nov 24 09:10:55 crc kubenswrapper[4944]: I1124 09:10:55.931336 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5kzx5"] Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.016951 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-79qbw"] Nov 24 09:10:56 crc kubenswrapper[4944]: E1124 09:10:56.017396 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2d5a94f-767f-4203-bf57-9172eccdce70" containerName="init" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.017417 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2d5a94f-767f-4203-bf57-9172eccdce70" containerName="init" Nov 24 09:10:56 crc kubenswrapper[4944]: E1124 09:10:56.017453 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af61f51-5798-4a0c-83a0-6dfd6d20d96a" containerName="keystone-bootstrap" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.017461 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af61f51-5798-4a0c-83a0-6dfd6d20d96a" containerName="keystone-bootstrap" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.018289 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2d5a94f-767f-4203-bf57-9172eccdce70" containerName="init" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.018363 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af61f51-5798-4a0c-83a0-6dfd6d20d96a" containerName="keystone-bootstrap" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.019170 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.021674 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.021951 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.022174 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qztjd" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.022358 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.023579 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.031466 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-79qbw"] Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.075908 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-credential-keys\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.076359 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-combined-ca-bundle\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.076578 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slt6l\" (UniqueName: \"kubernetes.io/projected/1d229255-c311-4626-be89-4769eca20010-kube-api-access-slt6l\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.077221 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-fernet-keys\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.077332 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-config-data\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.077361 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-scripts\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.179756 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-config-data\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.179813 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-scripts\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.179860 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-credential-keys\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.179899 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-combined-ca-bundle\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.179949 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slt6l\" (UniqueName: \"kubernetes.io/projected/1d229255-c311-4626-be89-4769eca20010-kube-api-access-slt6l\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.180074 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-fernet-keys\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.185115 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-credential-keys\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.186322 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-combined-ca-bundle\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.187351 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-config-data\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.193728 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-scripts\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.195961 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-fernet-keys\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.203924 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slt6l\" (UniqueName: \"kubernetes.io/projected/1d229255-c311-4626-be89-4769eca20010-kube-api-access-slt6l\") pod \"keystone-bootstrap-79qbw\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.288175 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5af61f51-5798-4a0c-83a0-6dfd6d20d96a" path="/var/lib/kubelet/pods/5af61f51-5798-4a0c-83a0-6dfd6d20d96a/volumes" Nov 24 09:10:56 crc kubenswrapper[4944]: I1124 09:10:56.345988 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.119460 4944 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.120331 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tt4k7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-hpjlj_openstack(6c338ca6-2795-4c94-b863-23579147b584): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.121786 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-hpjlj" podUID="6c338ca6-2795-4c94-b863-23579147b584" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.178492 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.190778 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.211544 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251536 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-combined-ca-bundle\") pod \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251581 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-scripts\") pod \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251619 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-httpd-run\") pod \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251641 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-config-data\") pod \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251678 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-scripts\") pod \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251707 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-config-data\") pod \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251726 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251785 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-public-tls-certs\") pod \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251861 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-logs\") pod \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251918 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251934 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-internal-tls-certs\") pod \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251957 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxtjh\" (UniqueName: \"kubernetes.io/projected/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-kube-api-access-dxtjh\") pod \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.251990 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-httpd-run\") pod \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.252023 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlfnr\" (UniqueName: \"kubernetes.io/projected/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-kube-api-access-tlfnr\") pod \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.252051 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-combined-ca-bundle\") pod \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\" (UID: \"9a0f99e5-5705-40dd-bc57-8faf9deb0aad\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.252109 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-logs\") pod \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\" (UID: \"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.253306 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-logs" (OuterVolumeSpecName: "logs") pod "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" (UID: "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.253645 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9a0f99e5-5705-40dd-bc57-8faf9deb0aad" (UID: "9a0f99e5-5705-40dd-bc57-8faf9deb0aad"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.257500 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-scripts" (OuterVolumeSpecName: "scripts") pod "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" (UID: "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.259282 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" (UID: "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.259579 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-logs" (OuterVolumeSpecName: "logs") pod "9a0f99e5-5705-40dd-bc57-8faf9deb0aad" (UID: "9a0f99e5-5705-40dd-bc57-8faf9deb0aad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.260089 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-kube-api-access-tlfnr" (OuterVolumeSpecName: "kube-api-access-tlfnr") pod "9a0f99e5-5705-40dd-bc57-8faf9deb0aad" (UID: "9a0f99e5-5705-40dd-bc57-8faf9deb0aad"). InnerVolumeSpecName "kube-api-access-tlfnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.260608 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "9a0f99e5-5705-40dd-bc57-8faf9deb0aad" (UID: "9a0f99e5-5705-40dd-bc57-8faf9deb0aad"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.261238 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" (UID: "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.262318 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-kube-api-access-dxtjh" (OuterVolumeSpecName: "kube-api-access-dxtjh") pod "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" (UID: "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd"). InnerVolumeSpecName "kube-api-access-dxtjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.275825 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-scripts" (OuterVolumeSpecName: "scripts") pod "9a0f99e5-5705-40dd-bc57-8faf9deb0aad" (UID: "9a0f99e5-5705-40dd-bc57-8faf9deb0aad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.301192 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a0f99e5-5705-40dd-bc57-8faf9deb0aad" (UID: "9a0f99e5-5705-40dd-bc57-8faf9deb0aad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.304901 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" (UID: "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.319642 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" (UID: "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.322035 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9a0f99e5-5705-40dd-bc57-8faf9deb0aad" (UID: "9a0f99e5-5705-40dd-bc57-8faf9deb0aad"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.323122 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-config-data" (OuterVolumeSpecName: "config-data") pod "9a0f99e5-5705-40dd-bc57-8faf9deb0aad" (UID: "9a0f99e5-5705-40dd-bc57-8faf9deb0aad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.335833 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-config-data" (OuterVolumeSpecName: "config-data") pod "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" (UID: "f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.354008 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-combined-ca-bundle\") pod \"587209cc-b2bb-460e-a6d4-1391b6294864\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.354092 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-config\") pod \"587209cc-b2bb-460e-a6d4-1391b6294864\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.354621 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x54mf\" (UniqueName: \"kubernetes.io/projected/587209cc-b2bb-460e-a6d4-1391b6294864-kube-api-access-x54mf\") pod \"587209cc-b2bb-460e-a6d4-1391b6294864\" (UID: \"587209cc-b2bb-460e-a6d4-1391b6294864\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355147 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355169 4944 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355178 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxtjh\" (UniqueName: \"kubernetes.io/projected/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-kube-api-access-dxtjh\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355187 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355195 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlfnr\" (UniqueName: \"kubernetes.io/projected/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-kube-api-access-tlfnr\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355203 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355212 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355220 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355236 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355696 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355713 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355722 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355730 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355747 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355755 4944 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.355763 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a0f99e5-5705-40dd-bc57-8faf9deb0aad-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.360354 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/587209cc-b2bb-460e-a6d4-1391b6294864-kube-api-access-x54mf" (OuterVolumeSpecName: "kube-api-access-x54mf") pod "587209cc-b2bb-460e-a6d4-1391b6294864" (UID: "587209cc-b2bb-460e-a6d4-1391b6294864"). InnerVolumeSpecName "kube-api-access-x54mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.380510 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.382415 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "587209cc-b2bb-460e-a6d4-1391b6294864" (UID: "587209cc-b2bb-460e-a6d4-1391b6294864"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.385079 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.386400 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-config" (OuterVolumeSpecName: "config") pod "587209cc-b2bb-460e-a6d4-1391b6294864" (UID: "587209cc-b2bb-460e-a6d4-1391b6294864"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.456940 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.456976 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.456989 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/587209cc-b2bb-460e-a6d4-1391b6294864-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.457001 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.457014 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x54mf\" (UniqueName: \"kubernetes.io/projected/587209cc-b2bb-460e-a6d4-1391b6294864-kube-api-access-x54mf\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.673607 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.673688 4944 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.673869 4944 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mhz6v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-nst6k_openstack(0493ffa7-e203-4f40-b117-07b07fe13380): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.675198 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-nst6k" podUID="0493ffa7-e203-4f40-b117-07b07fe13380" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.754964 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" event={"ID":"fc2988bc-f594-4018-b585-67800dc75b9b","Type":"ContainerDied","Data":"99b48245f2018fca11fde69aa3674bd6276a03636b114739e54a8c0bc8c2d3fb"} Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.755239 4944 scope.go:117] "RemoveContainer" containerID="dad14e323f57ede01cb229731af038ba2f3cae867b2341ccd19c185119305a67" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.755409 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-4b5pb" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.760939 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-config\") pod \"fc2988bc-f594-4018-b585-67800dc75b9b\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.761022 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-nb\") pod \"fc2988bc-f594-4018-b585-67800dc75b9b\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.761104 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-swift-storage-0\") pod \"fc2988bc-f594-4018-b585-67800dc75b9b\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.761398 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-sb\") pod \"fc2988bc-f594-4018-b585-67800dc75b9b\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.761438 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk4lz\" (UniqueName: \"kubernetes.io/projected/fc2988bc-f594-4018-b585-67800dc75b9b-kube-api-access-sk4lz\") pod \"fc2988bc-f594-4018-b585-67800dc75b9b\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.761442 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9a0f99e5-5705-40dd-bc57-8faf9deb0aad","Type":"ContainerDied","Data":"59a930b0ce32c8073992fb22af39476cb95da8680827f64832667159f9ddb983"} Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.761493 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-svc\") pod \"fc2988bc-f594-4018-b585-67800dc75b9b\" (UID: \"fc2988bc-f594-4018-b585-67800dc75b9b\") " Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.761534 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.766762 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc2988bc-f594-4018-b585-67800dc75b9b-kube-api-access-sk4lz" (OuterVolumeSpecName: "kube-api-access-sk4lz") pod "fc2988bc-f594-4018-b585-67800dc75b9b" (UID: "fc2988bc-f594-4018-b585-67800dc75b9b"). InnerVolumeSpecName "kube-api-access-sk4lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.776164 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vqngq" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.776175 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vqngq" event={"ID":"587209cc-b2bb-460e-a6d4-1391b6294864","Type":"ContainerDied","Data":"31e43085dd5c27d1da92b39e70a0dfd51e3a7d222fa701ed81e352d17778db0a"} Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.776211 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31e43085dd5c27d1da92b39e70a0dfd51e3a7d222fa701ed81e352d17778db0a" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.780099 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.782256 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd","Type":"ContainerDied","Data":"262452f701b1841f46f9ae9fc5c06156b54eca41f5c73a3e87476dfd93624ba9"} Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.784251 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-nst6k" podUID="0493ffa7-e203-4f40-b117-07b07fe13380" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.789028 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-hpjlj" podUID="6c338ca6-2795-4c94-b863-23579147b584" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.802836 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.822780 4944 scope.go:117] "RemoveContainer" containerID="650775d8fb82494dc6f1f67ef4f93ee0286d62e5144131ad3ffafef35bd52569" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.822899 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.838178 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.838642 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2988bc-f594-4018-b585-67800dc75b9b" containerName="dnsmasq-dns" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.838666 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2988bc-f594-4018-b585-67800dc75b9b" containerName="dnsmasq-dns" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.838682 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0f99e5-5705-40dd-bc57-8faf9deb0aad" containerName="glance-log" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.838691 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0f99e5-5705-40dd-bc57-8faf9deb0aad" containerName="glance-log" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.838703 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2988bc-f594-4018-b585-67800dc75b9b" containerName="init" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.838711 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2988bc-f594-4018-b585-67800dc75b9b" containerName="init" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.838730 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" containerName="glance-httpd" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.838745 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" containerName="glance-httpd" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.838764 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" containerName="glance-log" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.838772 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" containerName="glance-log" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.838789 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="587209cc-b2bb-460e-a6d4-1391b6294864" containerName="neutron-db-sync" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.838796 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="587209cc-b2bb-460e-a6d4-1391b6294864" containerName="neutron-db-sync" Nov 24 09:10:58 crc kubenswrapper[4944]: E1124 09:10:58.838814 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0f99e5-5705-40dd-bc57-8faf9deb0aad" containerName="glance-httpd" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.838821 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0f99e5-5705-40dd-bc57-8faf9deb0aad" containerName="glance-httpd" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.839197 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0f99e5-5705-40dd-bc57-8faf9deb0aad" containerName="glance-log" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.839214 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" containerName="glance-log" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.839225 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" containerName="glance-httpd" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.839236 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0f99e5-5705-40dd-bc57-8faf9deb0aad" containerName="glance-httpd" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.839251 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="587209cc-b2bb-460e-a6d4-1391b6294864" containerName="neutron-db-sync" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.839266 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc2988bc-f594-4018-b585-67800dc75b9b" containerName="dnsmasq-dns" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.842525 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.846516 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.847262 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.847438 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.847582 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-9rvqc" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.849108 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.867593 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk4lz\" (UniqueName: \"kubernetes.io/projected/fc2988bc-f594-4018-b585-67800dc75b9b-kube-api-access-sk4lz\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.893882 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.915342 4944 scope.go:117] "RemoveContainer" containerID="c71680738e1c9564a89a6810520284a64b66da8b6f0a69ee9842305d14ac903c" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.920256 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.932204 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.934125 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.940897 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.941173 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.954567 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.960630 4944 scope.go:117] "RemoveContainer" containerID="f8955f4029dd931e250c21461cd99c9ce093999fbaf9a8b7c8d006905e56b1a9" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.973664 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-scripts\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.973750 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-logs\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.973921 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.973980 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.974006 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-config-data\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.974031 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.974078 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc86b\" (UniqueName: \"kubernetes.io/projected/1221a90c-370a-4103-bd3c-b3dfe670b446-kube-api-access-lc86b\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:58 crc kubenswrapper[4944]: I1124 09:10:58.974145 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.000727 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fc2988bc-f594-4018-b585-67800dc75b9b" (UID: "fc2988bc-f594-4018-b585-67800dc75b9b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.008279 4944 scope.go:117] "RemoveContainer" containerID="440f5c50d132372c6d902d2a2f00660b621fb19a9c9e8368dd186c043e190d94" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.013046 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fc2988bc-f594-4018-b585-67800dc75b9b" (UID: "fc2988bc-f594-4018-b585-67800dc75b9b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.024200 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fc2988bc-f594-4018-b585-67800dc75b9b" (UID: "fc2988bc-f594-4018-b585-67800dc75b9b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.047319 4944 scope.go:117] "RemoveContainer" containerID="03e2a6129f9e35125553e7b1098cb7459e8a4daaad033636c72180b697f6d667" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.051980 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-config" (OuterVolumeSpecName: "config") pod "fc2988bc-f594-4018-b585-67800dc75b9b" (UID: "fc2988bc-f594-4018-b585-67800dc75b9b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.056558 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fc2988bc-f594-4018-b585-67800dc75b9b" (UID: "fc2988bc-f594-4018-b585-67800dc75b9b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077165 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-scripts\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077216 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-logs\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077255 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077293 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vmwp\" (UniqueName: \"kubernetes.io/projected/89e12262-3506-49db-b83c-e9a53974ad5d-kube-api-access-4vmwp\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077325 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077366 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077394 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077427 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077448 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077470 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077489 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-logs\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077511 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-config-data\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077535 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077554 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc86b\" (UniqueName: \"kubernetes.io/projected/1221a90c-370a-4103-bd3c-b3dfe670b446-kube-api-access-lc86b\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077580 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077601 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077642 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077652 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077662 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077671 4944 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077680 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc2988bc-f594-4018-b585-67800dc75b9b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.077933 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-logs\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.078154 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.078429 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.082665 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-scripts\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.083198 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-config-data\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.092982 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.093117 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.101656 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc86b\" (UniqueName: \"kubernetes.io/projected/1221a90c-370a-4103-bd3c-b3dfe670b446-kube-api-access-lc86b\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.104634 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.181094 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vmwp\" (UniqueName: \"kubernetes.io/projected/89e12262-3506-49db-b83c-e9a53974ad5d-kube-api-access-4vmwp\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.181172 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.181210 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.181259 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.181288 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.181329 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-logs\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.181396 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.181489 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.181728 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.182037 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.185108 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-logs\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.186603 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.188022 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.189221 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.193400 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.224639 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.228358 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vmwp\" (UniqueName: \"kubernetes.io/projected/89e12262-3506-49db-b83c-e9a53974ad5d-kube-api-access-4vmwp\") pod \"glance-default-internal-api-0\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.244879 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.254899 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-79qbw"] Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.292914 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.423737 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-4b5pb"] Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.447233 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-4b5pb"] Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.468306 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-795tt"] Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.476305 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.499032 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-795tt"] Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.577904 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-54f69f9546-4ksss"] Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.580740 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.586732 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54f69f9546-4ksss"] Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.599881 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.600223 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.600233 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.600310 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.600355 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-svc\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.600440 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8msfp\" (UniqueName: \"kubernetes.io/projected/7b35c8ac-158b-480d-9dcc-d0a474151efa-kube-api-access-8msfp\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.600464 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-config\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.600533 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.600369 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-vk9ck" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.600407 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.701744 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-combined-ca-bundle\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.702199 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.702242 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-config\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.702272 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.702302 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhpsd\" (UniqueName: \"kubernetes.io/projected/0c922b8a-f6f5-4801-839c-e67fcef0a67b-kube-api-access-qhpsd\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.702332 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-svc\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.702396 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8msfp\" (UniqueName: \"kubernetes.io/projected/7b35c8ac-158b-480d-9dcc-d0a474151efa-kube-api-access-8msfp\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.702422 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-config\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.702462 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-ovndb-tls-certs\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.702489 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-httpd-config\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.702516 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.703673 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.704361 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.704936 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.707122 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-svc\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.707714 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-config\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.738080 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8msfp\" (UniqueName: \"kubernetes.io/projected/7b35c8ac-158b-480d-9dcc-d0a474151efa-kube-api-access-8msfp\") pod \"dnsmasq-dns-6b7b667979-795tt\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.797443 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s7k5s" event={"ID":"6b3396e1-991a-4a6e-a7a6-665658af7e63","Type":"ContainerStarted","Data":"ceafa9e1a0d2708df2e25915e16d1a81cfc2d983d0d29259d0153aeb04d1d53f"} Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.803875 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-combined-ca-bundle\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.805641 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-config\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.805892 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhpsd\" (UniqueName: \"kubernetes.io/projected/0c922b8a-f6f5-4801-839c-e67fcef0a67b-kube-api-access-qhpsd\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.806194 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-ovndb-tls-certs\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.806314 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-httpd-config\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.807746 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-combined-ca-bundle\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.810849 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-config\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.811377 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-ovndb-tls-certs\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.813111 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-httpd-config\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.815214 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de53ccf2-83b3-4116-a8eb-d5ac212e11e7","Type":"ContainerStarted","Data":"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc"} Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.826540 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-79qbw" event={"ID":"1d229255-c311-4626-be89-4769eca20010","Type":"ContainerStarted","Data":"446e72ed669c7790b5e97ed81cfd34742c7c9da15ec2b51504d0e8b3f86995c1"} Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.826609 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-79qbw" event={"ID":"1d229255-c311-4626-be89-4769eca20010","Type":"ContainerStarted","Data":"d456a5d02a57a09303011d9c1620b3548bf6d1a1a1c89cd8430fff103dd16b29"} Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.830871 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-s7k5s" podStartSLOduration=2.28262487 podStartE2EDuration="23.830853364s" podCreationTimestamp="2025-11-24 09:10:36 +0000 UTC" firstStartedPulling="2025-11-24 09:10:37.130314304 +0000 UTC m=+1097.664754766" lastFinishedPulling="2025-11-24 09:10:58.678542798 +0000 UTC m=+1119.212983260" observedRunningTime="2025-11-24 09:10:59.819165261 +0000 UTC m=+1120.353605723" watchObservedRunningTime="2025-11-24 09:10:59.830853364 +0000 UTC m=+1120.365293826" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.833835 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhpsd\" (UniqueName: \"kubernetes.io/projected/0c922b8a-f6f5-4801-839c-e67fcef0a67b-kube-api-access-qhpsd\") pod \"neutron-54f69f9546-4ksss\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.846941 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-79qbw" podStartSLOduration=4.846923906 podStartE2EDuration="4.846923906s" podCreationTimestamp="2025-11-24 09:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:10:59.842002819 +0000 UTC m=+1120.376443291" watchObservedRunningTime="2025-11-24 09:10:59.846923906 +0000 UTC m=+1120.381364368" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.884163 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:10:59 crc kubenswrapper[4944]: I1124 09:10:59.928495 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:11:00 crc kubenswrapper[4944]: I1124 09:11:00.111563 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:11:00 crc kubenswrapper[4944]: I1124 09:11:00.318634 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a0f99e5-5705-40dd-bc57-8faf9deb0aad" path="/var/lib/kubelet/pods/9a0f99e5-5705-40dd-bc57-8faf9deb0aad/volumes" Nov 24 09:11:00 crc kubenswrapper[4944]: I1124 09:11:00.319977 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd" path="/var/lib/kubelet/pods/f5d8a3cd-632d-4ab2-9555-1dcac54dd7cd/volumes" Nov 24 09:11:00 crc kubenswrapper[4944]: I1124 09:11:00.320705 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc2988bc-f594-4018-b585-67800dc75b9b" path="/var/lib/kubelet/pods/fc2988bc-f594-4018-b585-67800dc75b9b/volumes" Nov 24 09:11:00 crc kubenswrapper[4944]: I1124 09:11:00.498487 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-795tt"] Nov 24 09:11:00 crc kubenswrapper[4944]: W1124 09:11:00.518353 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b35c8ac_158b_480d_9dcc_d0a474151efa.slice/crio-720a0bf11e716b8b0e59df8f177d4d811679f30ad0cb3124b26ab35aef17d04c WatchSource:0}: Error finding container 720a0bf11e716b8b0e59df8f177d4d811679f30ad0cb3124b26ab35aef17d04c: Status 404 returned error can't find the container with id 720a0bf11e716b8b0e59df8f177d4d811679f30ad0cb3124b26ab35aef17d04c Nov 24 09:11:00 crc kubenswrapper[4944]: I1124 09:11:00.762820 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54f69f9546-4ksss"] Nov 24 09:11:00 crc kubenswrapper[4944]: I1124 09:11:00.906690 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1221a90c-370a-4103-bd3c-b3dfe670b446","Type":"ContainerStarted","Data":"5fd55c50b1f605f2526663a959c2eef7286669d0dc1d75b76ac68a84af2bc045"} Nov 24 09:11:00 crc kubenswrapper[4944]: I1124 09:11:00.917863 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-795tt" event={"ID":"7b35c8ac-158b-480d-9dcc-d0a474151efa","Type":"ContainerStarted","Data":"720a0bf11e716b8b0e59df8f177d4d811679f30ad0cb3124b26ab35aef17d04c"} Nov 24 09:11:00 crc kubenswrapper[4944]: I1124 09:11:00.930829 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:11:01 crc kubenswrapper[4944]: W1124 09:11:01.401086 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c922b8a_f6f5_4801_839c_e67fcef0a67b.slice/crio-d0c3fe174959deb0066b04993bb241792858eb342200ba798d19f2f50caaa54e WatchSource:0}: Error finding container d0c3fe174959deb0066b04993bb241792858eb342200ba798d19f2f50caaa54e: Status 404 returned error can't find the container with id d0c3fe174959deb0066b04993bb241792858eb342200ba798d19f2f50caaa54e Nov 24 09:11:01 crc kubenswrapper[4944]: W1124 09:11:01.405733 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89e12262_3506_49db_b83c_e9a53974ad5d.slice/crio-006a23afeae96526189b1fd3964137c2676694fcbc430fbb0cd51846186df0d3 WatchSource:0}: Error finding container 006a23afeae96526189b1fd3964137c2676694fcbc430fbb0cd51846186df0d3: Status 404 returned error can't find the container with id 006a23afeae96526189b1fd3964137c2676694fcbc430fbb0cd51846186df0d3 Nov 24 09:11:01 crc kubenswrapper[4944]: I1124 09:11:01.936437 4944 generic.go:334] "Generic (PLEG): container finished" podID="7b35c8ac-158b-480d-9dcc-d0a474151efa" containerID="7bbaef1f0abf82ee381df9b478d9430d56e85b59230e0d8ecb5f76b4d607d961" exitCode=0 Nov 24 09:11:01 crc kubenswrapper[4944]: I1124 09:11:01.936597 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-795tt" event={"ID":"7b35c8ac-158b-480d-9dcc-d0a474151efa","Type":"ContainerDied","Data":"7bbaef1f0abf82ee381df9b478d9430d56e85b59230e0d8ecb5f76b4d607d961"} Nov 24 09:11:01 crc kubenswrapper[4944]: I1124 09:11:01.947285 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de53ccf2-83b3-4116-a8eb-d5ac212e11e7","Type":"ContainerStarted","Data":"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c"} Nov 24 09:11:01 crc kubenswrapper[4944]: I1124 09:11:01.954540 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1221a90c-370a-4103-bd3c-b3dfe670b446","Type":"ContainerStarted","Data":"0d336c6d00011b405b89e69c75750bd4892dffa8ff7440cdf41f4517d10dd129"} Nov 24 09:11:01 crc kubenswrapper[4944]: I1124 09:11:01.955766 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f69f9546-4ksss" event={"ID":"0c922b8a-f6f5-4801-839c-e67fcef0a67b","Type":"ContainerStarted","Data":"02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec"} Nov 24 09:11:01 crc kubenswrapper[4944]: I1124 09:11:01.955790 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f69f9546-4ksss" event={"ID":"0c922b8a-f6f5-4801-839c-e67fcef0a67b","Type":"ContainerStarted","Data":"d0c3fe174959deb0066b04993bb241792858eb342200ba798d19f2f50caaa54e"} Nov 24 09:11:01 crc kubenswrapper[4944]: I1124 09:11:01.964767 4944 generic.go:334] "Generic (PLEG): container finished" podID="6b3396e1-991a-4a6e-a7a6-665658af7e63" containerID="ceafa9e1a0d2708df2e25915e16d1a81cfc2d983d0d29259d0153aeb04d1d53f" exitCode=0 Nov 24 09:11:01 crc kubenswrapper[4944]: I1124 09:11:01.964864 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s7k5s" event={"ID":"6b3396e1-991a-4a6e-a7a6-665658af7e63","Type":"ContainerDied","Data":"ceafa9e1a0d2708df2e25915e16d1a81cfc2d983d0d29259d0153aeb04d1d53f"} Nov 24 09:11:01 crc kubenswrapper[4944]: I1124 09:11:01.977500 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"89e12262-3506-49db-b83c-e9a53974ad5d","Type":"ContainerStarted","Data":"006a23afeae96526189b1fd3964137c2676694fcbc430fbb0cd51846186df0d3"} Nov 24 09:11:02 crc kubenswrapper[4944]: I1124 09:11:02.991715 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-795tt" event={"ID":"7b35c8ac-158b-480d-9dcc-d0a474151efa","Type":"ContainerStarted","Data":"b9dc3d12c1db828a9cd9ce34593a1437825bdb491b2df896b66efac37ce2a267"} Nov 24 09:11:02 crc kubenswrapper[4944]: I1124 09:11:02.992353 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.001563 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1221a90c-370a-4103-bd3c-b3dfe670b446","Type":"ContainerStarted","Data":"32fb6228e84968d5e793ee6c46d782e11a2ef4eebf6c818f2226f27364162b50"} Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.008708 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f69f9546-4ksss" event={"ID":"0c922b8a-f6f5-4801-839c-e67fcef0a67b","Type":"ContainerStarted","Data":"82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a"} Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.008940 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.012186 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"89e12262-3506-49db-b83c-e9a53974ad5d","Type":"ContainerStarted","Data":"8bda4061dc19bf95a403f1262f4d42e018ccabe08f270939e7945ba45cb194b9"} Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.031405 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-795tt" podStartSLOduration=4.03138069 podStartE2EDuration="4.03138069s" podCreationTimestamp="2025-11-24 09:10:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:03.016527866 +0000 UTC m=+1123.550968348" watchObservedRunningTime="2025-11-24 09:11:03.03138069 +0000 UTC m=+1123.565821152" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.043987 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-54f69f9546-4ksss" podStartSLOduration=4.04397051 podStartE2EDuration="4.04397051s" podCreationTimestamp="2025-11-24 09:10:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:03.042506394 +0000 UTC m=+1123.576946856" watchObservedRunningTime="2025-11-24 09:11:03.04397051 +0000 UTC m=+1123.578410972" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.081465 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.081435195 podStartE2EDuration="5.081435195s" podCreationTimestamp="2025-11-24 09:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:03.079591615 +0000 UTC m=+1123.614032087" watchObservedRunningTime="2025-11-24 09:11:03.081435195 +0000 UTC m=+1123.615875657" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.425638 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s7k5s" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.505151 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-scripts\") pod \"6b3396e1-991a-4a6e-a7a6-665658af7e63\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.505276 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29dcx\" (UniqueName: \"kubernetes.io/projected/6b3396e1-991a-4a6e-a7a6-665658af7e63-kube-api-access-29dcx\") pod \"6b3396e1-991a-4a6e-a7a6-665658af7e63\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.505340 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-combined-ca-bundle\") pod \"6b3396e1-991a-4a6e-a7a6-665658af7e63\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.505544 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b3396e1-991a-4a6e-a7a6-665658af7e63-logs\") pod \"6b3396e1-991a-4a6e-a7a6-665658af7e63\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.505726 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-config-data\") pod \"6b3396e1-991a-4a6e-a7a6-665658af7e63\" (UID: \"6b3396e1-991a-4a6e-a7a6-665658af7e63\") " Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.509248 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b3396e1-991a-4a6e-a7a6-665658af7e63-logs" (OuterVolumeSpecName: "logs") pod "6b3396e1-991a-4a6e-a7a6-665658af7e63" (UID: "6b3396e1-991a-4a6e-a7a6-665658af7e63"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.514983 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b3396e1-991a-4a6e-a7a6-665658af7e63-kube-api-access-29dcx" (OuterVolumeSpecName: "kube-api-access-29dcx") pod "6b3396e1-991a-4a6e-a7a6-665658af7e63" (UID: "6b3396e1-991a-4a6e-a7a6-665658af7e63"). InnerVolumeSpecName "kube-api-access-29dcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.523286 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-scripts" (OuterVolumeSpecName: "scripts") pod "6b3396e1-991a-4a6e-a7a6-665658af7e63" (UID: "6b3396e1-991a-4a6e-a7a6-665658af7e63"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.555613 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b3396e1-991a-4a6e-a7a6-665658af7e63" (UID: "6b3396e1-991a-4a6e-a7a6-665658af7e63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.578657 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-config-data" (OuterVolumeSpecName: "config-data") pod "6b3396e1-991a-4a6e-a7a6-665658af7e63" (UID: "6b3396e1-991a-4a6e-a7a6-665658af7e63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.609220 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b3396e1-991a-4a6e-a7a6-665658af7e63-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.609276 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.609290 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.609308 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29dcx\" (UniqueName: \"kubernetes.io/projected/6b3396e1-991a-4a6e-a7a6-665658af7e63-kube-api-access-29dcx\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.609323 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3396e1-991a-4a6e-a7a6-665658af7e63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.639887 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d5d955fb7-kf4q4"] Nov 24 09:11:03 crc kubenswrapper[4944]: E1124 09:11:03.651476 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b3396e1-991a-4a6e-a7a6-665658af7e63" containerName="placement-db-sync" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.651608 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b3396e1-991a-4a6e-a7a6-665658af7e63" containerName="placement-db-sync" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.651905 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b3396e1-991a-4a6e-a7a6-665658af7e63" containerName="placement-db-sync" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.653073 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.656903 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.657159 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.657584 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d5d955fb7-kf4q4"] Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.710998 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-combined-ca-bundle\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.711096 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st78n\" (UniqueName: \"kubernetes.io/projected/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-kube-api-access-st78n\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.711130 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-ovndb-tls-certs\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.711153 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-public-tls-certs\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.711392 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-httpd-config\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.711450 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-internal-tls-certs\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.711744 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-config\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.813972 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-httpd-config\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.814070 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-internal-tls-certs\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.814132 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-config\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.814216 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-combined-ca-bundle\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.814238 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st78n\" (UniqueName: \"kubernetes.io/projected/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-kube-api-access-st78n\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.814262 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-ovndb-tls-certs\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.814282 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-public-tls-certs\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.829213 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-httpd-config\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.833666 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-internal-tls-certs\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.835704 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-public-tls-certs\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.837576 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-combined-ca-bundle\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.837756 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-ovndb-tls-certs\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.838187 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-config\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:03 crc kubenswrapper[4944]: I1124 09:11:03.880325 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st78n\" (UniqueName: \"kubernetes.io/projected/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-kube-api-access-st78n\") pod \"neutron-5d5d955fb7-kf4q4\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.026013 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.031092 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s7k5s" event={"ID":"6b3396e1-991a-4a6e-a7a6-665658af7e63","Type":"ContainerDied","Data":"a6e037ef5c3b909e2f83c757deaa7e41ad0c44c422292fc6fca384dfab8f39e4"} Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.031234 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6e037ef5c3b909e2f83c757deaa7e41ad0c44c422292fc6fca384dfab8f39e4" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.031105 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s7k5s" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.038893 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"89e12262-3506-49db-b83c-e9a53974ad5d","Type":"ContainerStarted","Data":"d3c3b5c16d1ed60f1fdf53659af98f1126eb3740de0da4f9ea843419e7af7ba9"} Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.047071 4944 generic.go:334] "Generic (PLEG): container finished" podID="1d229255-c311-4626-be89-4769eca20010" containerID="446e72ed669c7790b5e97ed81cfd34742c7c9da15ec2b51504d0e8b3f86995c1" exitCode=0 Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.047175 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-79qbw" event={"ID":"1d229255-c311-4626-be89-4769eca20010","Type":"ContainerDied","Data":"446e72ed669c7790b5e97ed81cfd34742c7c9da15ec2b51504d0e8b3f86995c1"} Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.229959 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5ccb7b4f9b-mcdmd"] Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.244078 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.247243 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.247857 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.247938 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9lk6q" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.247964 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.250225 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.279861 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5ccb7b4f9b-mcdmd"] Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.339838 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-config-data\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.340205 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-scripts\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.340252 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70a59d46-95bf-4681-b61d-22239c638737-logs\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.340483 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-internal-tls-certs\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.340596 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-public-tls-certs\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.340668 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-combined-ca-bundle\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.340688 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl4bh\" (UniqueName: \"kubernetes.io/projected/70a59d46-95bf-4681-b61d-22239c638737-kube-api-access-gl4bh\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.442268 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-config-data\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.442360 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-scripts\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.442392 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70a59d46-95bf-4681-b61d-22239c638737-logs\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.442465 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-internal-tls-certs\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.442494 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-public-tls-certs\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.442535 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-combined-ca-bundle\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.442558 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl4bh\" (UniqueName: \"kubernetes.io/projected/70a59d46-95bf-4681-b61d-22239c638737-kube-api-access-gl4bh\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.447590 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70a59d46-95bf-4681-b61d-22239c638737-logs\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.449636 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-scripts\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.449926 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-config-data\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.453865 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-public-tls-certs\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.457693 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-combined-ca-bundle\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.466971 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl4bh\" (UniqueName: \"kubernetes.io/projected/70a59d46-95bf-4681-b61d-22239c638737-kube-api-access-gl4bh\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.467014 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-internal-tls-certs\") pod \"placement-5ccb7b4f9b-mcdmd\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.577388 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:04 crc kubenswrapper[4944]: I1124 09:11:04.703652 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d5d955fb7-kf4q4"] Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.071931 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d5d955fb7-kf4q4" event={"ID":"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6","Type":"ContainerStarted","Data":"ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d"} Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.072263 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d5d955fb7-kf4q4" event={"ID":"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6","Type":"ContainerStarted","Data":"5b0f096028841e609c6de90a324e48f2efd19aa29e3d69de72b1f9c68f012de8"} Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.114283 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.114267205 podStartE2EDuration="7.114267205s" podCreationTimestamp="2025-11-24 09:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:05.106488626 +0000 UTC m=+1125.640929098" watchObservedRunningTime="2025-11-24 09:11:05.114267205 +0000 UTC m=+1125.648707667" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.127162 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5ccb7b4f9b-mcdmd"] Nov 24 09:11:05 crc kubenswrapper[4944]: W1124 09:11:05.136166 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70a59d46_95bf_4681_b61d_22239c638737.slice/crio-c3f8d5efd38123a3c9e348998d2cbf89144b5041be52f1cdb8f560f749d658c2 WatchSource:0}: Error finding container c3f8d5efd38123a3c9e348998d2cbf89144b5041be52f1cdb8f560f749d658c2: Status 404 returned error can't find the container with id c3f8d5efd38123a3c9e348998d2cbf89144b5041be52f1cdb8f560f749d658c2 Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.517018 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.676226 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-scripts\") pod \"1d229255-c311-4626-be89-4769eca20010\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.676295 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-credential-keys\") pod \"1d229255-c311-4626-be89-4769eca20010\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.676400 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slt6l\" (UniqueName: \"kubernetes.io/projected/1d229255-c311-4626-be89-4769eca20010-kube-api-access-slt6l\") pod \"1d229255-c311-4626-be89-4769eca20010\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.676425 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-fernet-keys\") pod \"1d229255-c311-4626-be89-4769eca20010\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.676532 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-combined-ca-bundle\") pod \"1d229255-c311-4626-be89-4769eca20010\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.676583 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-config-data\") pod \"1d229255-c311-4626-be89-4769eca20010\" (UID: \"1d229255-c311-4626-be89-4769eca20010\") " Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.682566 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d229255-c311-4626-be89-4769eca20010-kube-api-access-slt6l" (OuterVolumeSpecName: "kube-api-access-slt6l") pod "1d229255-c311-4626-be89-4769eca20010" (UID: "1d229255-c311-4626-be89-4769eca20010"). InnerVolumeSpecName "kube-api-access-slt6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.685924 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1d229255-c311-4626-be89-4769eca20010" (UID: "1d229255-c311-4626-be89-4769eca20010"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.687431 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-scripts" (OuterVolumeSpecName: "scripts") pod "1d229255-c311-4626-be89-4769eca20010" (UID: "1d229255-c311-4626-be89-4769eca20010"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.688367 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1d229255-c311-4626-be89-4769eca20010" (UID: "1d229255-c311-4626-be89-4769eca20010"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.713417 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d229255-c311-4626-be89-4769eca20010" (UID: "1d229255-c311-4626-be89-4769eca20010"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.723922 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-config-data" (OuterVolumeSpecName: "config-data") pod "1d229255-c311-4626-be89-4769eca20010" (UID: "1d229255-c311-4626-be89-4769eca20010"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.779183 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.779223 4944 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.779236 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slt6l\" (UniqueName: \"kubernetes.io/projected/1d229255-c311-4626-be89-4769eca20010-kube-api-access-slt6l\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.779248 4944 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.779262 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:05 crc kubenswrapper[4944]: I1124 09:11:05.779272 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d229255-c311-4626-be89-4769eca20010-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.080211 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5ccb7b4f9b-mcdmd" event={"ID":"70a59d46-95bf-4681-b61d-22239c638737","Type":"ContainerStarted","Data":"4ad8bef4c684b16347a1d2ba8be599c596216230dd5dd3d9c7c090945c16c531"} Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.080525 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5ccb7b4f9b-mcdmd" event={"ID":"70a59d46-95bf-4681-b61d-22239c638737","Type":"ContainerStarted","Data":"c3f8d5efd38123a3c9e348998d2cbf89144b5041be52f1cdb8f560f749d658c2"} Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.082832 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-79qbw" event={"ID":"1d229255-c311-4626-be89-4769eca20010","Type":"ContainerDied","Data":"d456a5d02a57a09303011d9c1620b3548bf6d1a1a1c89cd8430fff103dd16b29"} Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.082856 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-79qbw" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.082865 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d456a5d02a57a09303011d9c1620b3548bf6d1a1a1c89cd8430fff103dd16b29" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.084734 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d5d955fb7-kf4q4" event={"ID":"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6","Type":"ContainerStarted","Data":"dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992"} Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.084954 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.110252 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d5d955fb7-kf4q4" podStartSLOduration=3.110234818 podStartE2EDuration="3.110234818s" podCreationTimestamp="2025-11-24 09:11:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:06.108860483 +0000 UTC m=+1126.643300945" watchObservedRunningTime="2025-11-24 09:11:06.110234818 +0000 UTC m=+1126.644675280" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.173832 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5457b9857f-kg9gm"] Nov 24 09:11:06 crc kubenswrapper[4944]: E1124 09:11:06.174326 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d229255-c311-4626-be89-4769eca20010" containerName="keystone-bootstrap" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.174349 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d229255-c311-4626-be89-4769eca20010" containerName="keystone-bootstrap" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.174584 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d229255-c311-4626-be89-4769eca20010" containerName="keystone-bootstrap" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.176640 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.186787 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.186899 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.187019 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.187072 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qztjd" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.187278 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.187420 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.202852 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5457b9857f-kg9gm"] Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.289820 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-internal-tls-certs\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.289877 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdx75\" (UniqueName: \"kubernetes.io/projected/d2495f8a-6f01-47e6-beff-f0d7345967a7-kube-api-access-wdx75\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.289916 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-public-tls-certs\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.290023 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-scripts\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.290103 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-combined-ca-bundle\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.290313 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-fernet-keys\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.290350 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-config-data\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.290528 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-credential-keys\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.394707 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-fernet-keys\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.394782 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-config-data\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.394887 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-credential-keys\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.394968 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-internal-tls-certs\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.395022 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdx75\" (UniqueName: \"kubernetes.io/projected/d2495f8a-6f01-47e6-beff-f0d7345967a7-kube-api-access-wdx75\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.395167 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-public-tls-certs\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.395200 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-scripts\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.395247 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-combined-ca-bundle\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.399883 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-combined-ca-bundle\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.404125 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-fernet-keys\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.406412 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-internal-tls-certs\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.406926 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-credential-keys\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.413876 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-public-tls-certs\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.413922 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-config-data\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.420341 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-scripts\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.434034 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdx75\" (UniqueName: \"kubernetes.io/projected/d2495f8a-6f01-47e6-beff-f0d7345967a7-kube-api-access-wdx75\") pod \"keystone-5457b9857f-kg9gm\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:06 crc kubenswrapper[4944]: I1124 09:11:06.523745 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:09 crc kubenswrapper[4944]: I1124 09:11:09.245357 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 09:11:09 crc kubenswrapper[4944]: I1124 09:11:09.247063 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 09:11:09 crc kubenswrapper[4944]: I1124 09:11:09.293535 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 09:11:09 crc kubenswrapper[4944]: I1124 09:11:09.293592 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 09:11:09 crc kubenswrapper[4944]: I1124 09:11:09.311369 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 09:11:09 crc kubenswrapper[4944]: I1124 09:11:09.312614 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 09:11:09 crc kubenswrapper[4944]: I1124 09:11:09.325504 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 09:11:09 crc kubenswrapper[4944]: I1124 09:11:09.347285 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 09:11:09 crc kubenswrapper[4944]: I1124 09:11:09.886899 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:11:09 crc kubenswrapper[4944]: I1124 09:11:09.951531 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-62h9v"] Nov 24 09:11:09 crc kubenswrapper[4944]: I1124 09:11:09.964182 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" podUID="28e6918e-c11e-423e-996b-c43ee889155c" containerName="dnsmasq-dns" containerID="cri-o://4c6a05fbd86035dc80b848ecd34654a1e05f03fb73dc590f509e5bdbc7cb4bc2" gracePeriod=10 Nov 24 09:11:10 crc kubenswrapper[4944]: I1124 09:11:10.131614 4944 generic.go:334] "Generic (PLEG): container finished" podID="28e6918e-c11e-423e-996b-c43ee889155c" containerID="4c6a05fbd86035dc80b848ecd34654a1e05f03fb73dc590f509e5bdbc7cb4bc2" exitCode=0 Nov 24 09:11:10 crc kubenswrapper[4944]: I1124 09:11:10.132149 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" event={"ID":"28e6918e-c11e-423e-996b-c43ee889155c","Type":"ContainerDied","Data":"4c6a05fbd86035dc80b848ecd34654a1e05f03fb73dc590f509e5bdbc7cb4bc2"} Nov 24 09:11:10 crc kubenswrapper[4944]: I1124 09:11:10.132913 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 09:11:10 crc kubenswrapper[4944]: I1124 09:11:10.132932 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 09:11:10 crc kubenswrapper[4944]: I1124 09:11:10.132942 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 09:11:10 crc kubenswrapper[4944]: I1124 09:11:10.133085 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.187654 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" event={"ID":"28e6918e-c11e-423e-996b-c43ee889155c","Type":"ContainerDied","Data":"fdb2abf2a0aa8c58a0a0e177d6607a5e3a4cb820cf5d78281663a7c4ce885250"} Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.188280 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdb2abf2a0aa8c58a0a0e177d6607a5e3a4cb820cf5d78281663a7c4ce885250" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.191241 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.309785 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5457b9857f-kg9gm"] Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.318073 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-svc\") pod \"28e6918e-c11e-423e-996b-c43ee889155c\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.318202 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f6j9\" (UniqueName: \"kubernetes.io/projected/28e6918e-c11e-423e-996b-c43ee889155c-kube-api-access-9f6j9\") pod \"28e6918e-c11e-423e-996b-c43ee889155c\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.318222 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-nb\") pod \"28e6918e-c11e-423e-996b-c43ee889155c\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.318311 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-sb\") pod \"28e6918e-c11e-423e-996b-c43ee889155c\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.319559 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-swift-storage-0\") pod \"28e6918e-c11e-423e-996b-c43ee889155c\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.319916 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-config\") pod \"28e6918e-c11e-423e-996b-c43ee889155c\" (UID: \"28e6918e-c11e-423e-996b-c43ee889155c\") " Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.327213 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28e6918e-c11e-423e-996b-c43ee889155c-kube-api-access-9f6j9" (OuterVolumeSpecName: "kube-api-access-9f6j9") pod "28e6918e-c11e-423e-996b-c43ee889155c" (UID: "28e6918e-c11e-423e-996b-c43ee889155c"). InnerVolumeSpecName "kube-api-access-9f6j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.380737 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "28e6918e-c11e-423e-996b-c43ee889155c" (UID: "28e6918e-c11e-423e-996b-c43ee889155c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.381166 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "28e6918e-c11e-423e-996b-c43ee889155c" (UID: "28e6918e-c11e-423e-996b-c43ee889155c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.381778 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "28e6918e-c11e-423e-996b-c43ee889155c" (UID: "28e6918e-c11e-423e-996b-c43ee889155c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.417606 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "28e6918e-c11e-423e-996b-c43ee889155c" (UID: "28e6918e-c11e-423e-996b-c43ee889155c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.423663 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.423691 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f6j9\" (UniqueName: \"kubernetes.io/projected/28e6918e-c11e-423e-996b-c43ee889155c-kube-api-access-9f6j9\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.423702 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.423710 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.423719 4944 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.427826 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-config" (OuterVolumeSpecName: "config") pod "28e6918e-c11e-423e-996b-c43ee889155c" (UID: "28e6918e-c11e-423e-996b-c43ee889155c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:11:11 crc kubenswrapper[4944]: I1124 09:11:11.525100 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28e6918e-c11e-423e-996b-c43ee889155c-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.204262 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5ccb7b4f9b-mcdmd" event={"ID":"70a59d46-95bf-4681-b61d-22239c638737","Type":"ContainerStarted","Data":"e42a48dc5f9a748005eb91765bdfa38b2ad4b4f974b03b15a72a6b00023e47a2"} Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.205000 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.205340 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.207606 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5457b9857f-kg9gm" event={"ID":"d2495f8a-6f01-47e6-beff-f0d7345967a7","Type":"ContainerStarted","Data":"5950163d1a4608641fb113f50f6737d5436fc15138c0f58c9d9ec29a06eda578"} Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.207640 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5457b9857f-kg9gm" event={"ID":"d2495f8a-6f01-47e6-beff-f0d7345967a7","Type":"ContainerStarted","Data":"1ff67457d994ff3182e6c566098fc4eb751cdbf765fc343f619851795f6a8e7a"} Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.207726 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.209649 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.209736 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-62h9v" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.209688 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de53ccf2-83b3-4116-a8eb-d5ac212e11e7","Type":"ContainerStarted","Data":"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5"} Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.209813 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.209696 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.210022 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.227332 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5ccb7b4f9b-mcdmd" podStartSLOduration=8.227316457 podStartE2EDuration="8.227316457s" podCreationTimestamp="2025-11-24 09:11:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:12.224107485 +0000 UTC m=+1132.758547957" watchObservedRunningTime="2025-11-24 09:11:12.227316457 +0000 UTC m=+1132.761756919" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.250309 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5457b9857f-kg9gm" podStartSLOduration=6.25029292 podStartE2EDuration="6.25029292s" podCreationTimestamp="2025-11-24 09:11:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:12.245817937 +0000 UTC m=+1132.780258409" watchObservedRunningTime="2025-11-24 09:11:12.25029292 +0000 UTC m=+1132.784733382" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.356542 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-62h9v"] Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.369847 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-62h9v"] Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.880401 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.889099 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 09:11:12 crc kubenswrapper[4944]: I1124 09:11:12.988123 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 09:11:13 crc kubenswrapper[4944]: I1124 09:11:13.221383 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hpjlj" event={"ID":"6c338ca6-2795-4c94-b863-23579147b584","Type":"ContainerStarted","Data":"026e0e21f1a64afd79436eb0f96f0119e4d149466ea95e71c69fb8bd412bf0e1"} Nov 24 09:11:13 crc kubenswrapper[4944]: I1124 09:11:13.221401 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 09:11:13 crc kubenswrapper[4944]: I1124 09:11:13.245267 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-hpjlj" podStartSLOduration=3.057076524 podStartE2EDuration="37.24524777s" podCreationTimestamp="2025-11-24 09:10:36 +0000 UTC" firstStartedPulling="2025-11-24 09:10:37.322391637 +0000 UTC m=+1097.856832099" lastFinishedPulling="2025-11-24 09:11:11.510562883 +0000 UTC m=+1132.045003345" observedRunningTime="2025-11-24 09:11:13.237476253 +0000 UTC m=+1133.771916725" watchObservedRunningTime="2025-11-24 09:11:13.24524777 +0000 UTC m=+1133.779688222" Nov 24 09:11:13 crc kubenswrapper[4944]: I1124 09:11:13.332862 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 09:11:13 crc kubenswrapper[4944]: I1124 09:11:13.602294 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:14 crc kubenswrapper[4944]: I1124 09:11:14.289806 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28e6918e-c11e-423e-996b-c43ee889155c" path="/var/lib/kubelet/pods/28e6918e-c11e-423e-996b-c43ee889155c/volumes" Nov 24 09:11:15 crc kubenswrapper[4944]: I1124 09:11:15.254880 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nst6k" event={"ID":"0493ffa7-e203-4f40-b117-07b07fe13380","Type":"ContainerStarted","Data":"60c6a295891b9eabbf6aff3ae56e1707132a7e25f4e5417b47601279a7b69df1"} Nov 24 09:11:15 crc kubenswrapper[4944]: I1124 09:11:15.275667 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-nst6k" podStartSLOduration=3.13355493 podStartE2EDuration="40.275639972s" podCreationTimestamp="2025-11-24 09:10:35 +0000 UTC" firstStartedPulling="2025-11-24 09:10:36.859871686 +0000 UTC m=+1097.394312148" lastFinishedPulling="2025-11-24 09:11:14.001956728 +0000 UTC m=+1134.536397190" observedRunningTime="2025-11-24 09:11:15.270810078 +0000 UTC m=+1135.805250540" watchObservedRunningTime="2025-11-24 09:11:15.275639972 +0000 UTC m=+1135.810080434" Nov 24 09:11:18 crc kubenswrapper[4944]: I1124 09:11:18.282795 4944 generic.go:334] "Generic (PLEG): container finished" podID="6c338ca6-2795-4c94-b863-23579147b584" containerID="026e0e21f1a64afd79436eb0f96f0119e4d149466ea95e71c69fb8bd412bf0e1" exitCode=0 Nov 24 09:11:18 crc kubenswrapper[4944]: I1124 09:11:18.284862 4944 generic.go:334] "Generic (PLEG): container finished" podID="0493ffa7-e203-4f40-b117-07b07fe13380" containerID="60c6a295891b9eabbf6aff3ae56e1707132a7e25f4e5417b47601279a7b69df1" exitCode=0 Nov 24 09:11:18 crc kubenswrapper[4944]: I1124 09:11:18.286997 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hpjlj" event={"ID":"6c338ca6-2795-4c94-b863-23579147b584","Type":"ContainerDied","Data":"026e0e21f1a64afd79436eb0f96f0119e4d149466ea95e71c69fb8bd412bf0e1"} Nov 24 09:11:18 crc kubenswrapper[4944]: I1124 09:11:18.288169 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nst6k" event={"ID":"0493ffa7-e203-4f40-b117-07b07fe13380","Type":"ContainerDied","Data":"60c6a295891b9eabbf6aff3ae56e1707132a7e25f4e5417b47601279a7b69df1"} Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.295101 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de53ccf2-83b3-4116-a8eb-d5ac212e11e7","Type":"ContainerStarted","Data":"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a"} Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.295524 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="ceilometer-central-agent" containerID="cri-o://a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc" gracePeriod=30 Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.295569 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="sg-core" containerID="cri-o://3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5" gracePeriod=30 Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.295606 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="ceilometer-notification-agent" containerID="cri-o://4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c" gracePeriod=30 Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.295621 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="proxy-httpd" containerID="cri-o://c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a" gracePeriod=30 Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.332669 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.196528156 podStartE2EDuration="43.332648175s" podCreationTimestamp="2025-11-24 09:10:36 +0000 UTC" firstStartedPulling="2025-11-24 09:10:37.559120241 +0000 UTC m=+1098.093560703" lastFinishedPulling="2025-11-24 09:11:18.69524027 +0000 UTC m=+1139.229680722" observedRunningTime="2025-11-24 09:11:19.322245084 +0000 UTC m=+1139.856685556" watchObservedRunningTime="2025-11-24 09:11:19.332648175 +0000 UTC m=+1139.867088637" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.576903 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.663061 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nst6k" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.711661 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-combined-ca-bundle\") pod \"6c338ca6-2795-4c94-b863-23579147b584\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.711717 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-config-data\") pod \"6c338ca6-2795-4c94-b863-23579147b584\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.711772 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c338ca6-2795-4c94-b863-23579147b584-etc-machine-id\") pod \"6c338ca6-2795-4c94-b863-23579147b584\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.711845 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c338ca6-2795-4c94-b863-23579147b584-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6c338ca6-2795-4c94-b863-23579147b584" (UID: "6c338ca6-2795-4c94-b863-23579147b584"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.712163 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt4k7\" (UniqueName: \"kubernetes.io/projected/6c338ca6-2795-4c94-b863-23579147b584-kube-api-access-tt4k7\") pod \"6c338ca6-2795-4c94-b863-23579147b584\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.712308 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-scripts\") pod \"6c338ca6-2795-4c94-b863-23579147b584\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.712331 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-db-sync-config-data\") pod \"6c338ca6-2795-4c94-b863-23579147b584\" (UID: \"6c338ca6-2795-4c94-b863-23579147b584\") " Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.713014 4944 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c338ca6-2795-4c94-b863-23579147b584-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.721243 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-scripts" (OuterVolumeSpecName: "scripts") pod "6c338ca6-2795-4c94-b863-23579147b584" (UID: "6c338ca6-2795-4c94-b863-23579147b584"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.721231 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6c338ca6-2795-4c94-b863-23579147b584" (UID: "6c338ca6-2795-4c94-b863-23579147b584"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.721498 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c338ca6-2795-4c94-b863-23579147b584-kube-api-access-tt4k7" (OuterVolumeSpecName: "kube-api-access-tt4k7") pod "6c338ca6-2795-4c94-b863-23579147b584" (UID: "6c338ca6-2795-4c94-b863-23579147b584"). InnerVolumeSpecName "kube-api-access-tt4k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.741827 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c338ca6-2795-4c94-b863-23579147b584" (UID: "6c338ca6-2795-4c94-b863-23579147b584"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.763766 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-config-data" (OuterVolumeSpecName: "config-data") pod "6c338ca6-2795-4c94-b863-23579147b584" (UID: "6c338ca6-2795-4c94-b863-23579147b584"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.813830 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-combined-ca-bundle\") pod \"0493ffa7-e203-4f40-b117-07b07fe13380\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.814006 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhz6v\" (UniqueName: \"kubernetes.io/projected/0493ffa7-e203-4f40-b117-07b07fe13380-kube-api-access-mhz6v\") pod \"0493ffa7-e203-4f40-b117-07b07fe13380\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.814147 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-db-sync-config-data\") pod \"0493ffa7-e203-4f40-b117-07b07fe13380\" (UID: \"0493ffa7-e203-4f40-b117-07b07fe13380\") " Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.814770 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.814790 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.814799 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt4k7\" (UniqueName: \"kubernetes.io/projected/6c338ca6-2795-4c94-b863-23579147b584-kube-api-access-tt4k7\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.814808 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.814839 4944 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6c338ca6-2795-4c94-b863-23579147b584-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.817290 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0493ffa7-e203-4f40-b117-07b07fe13380-kube-api-access-mhz6v" (OuterVolumeSpecName: "kube-api-access-mhz6v") pod "0493ffa7-e203-4f40-b117-07b07fe13380" (UID: "0493ffa7-e203-4f40-b117-07b07fe13380"). InnerVolumeSpecName "kube-api-access-mhz6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.819003 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0493ffa7-e203-4f40-b117-07b07fe13380" (UID: "0493ffa7-e203-4f40-b117-07b07fe13380"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.837302 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0493ffa7-e203-4f40-b117-07b07fe13380" (UID: "0493ffa7-e203-4f40-b117-07b07fe13380"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.916352 4944 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.916588 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0493ffa7-e203-4f40-b117-07b07fe13380-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:19 crc kubenswrapper[4944]: I1124 09:11:19.916689 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhz6v\" (UniqueName: \"kubernetes.io/projected/0493ffa7-e203-4f40-b117-07b07fe13380-kube-api-access-mhz6v\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.282144 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.313804 4944 generic.go:334] "Generic (PLEG): container finished" podID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerID="c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a" exitCode=0 Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.313841 4944 generic.go:334] "Generic (PLEG): container finished" podID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerID="3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5" exitCode=2 Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.313852 4944 generic.go:334] "Generic (PLEG): container finished" podID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerID="4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c" exitCode=0 Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.313861 4944 generic.go:334] "Generic (PLEG): container finished" podID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerID="a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc" exitCode=0 Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.313917 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de53ccf2-83b3-4116-a8eb-d5ac212e11e7","Type":"ContainerDied","Data":"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a"} Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.313948 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de53ccf2-83b3-4116-a8eb-d5ac212e11e7","Type":"ContainerDied","Data":"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5"} Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.313962 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de53ccf2-83b3-4116-a8eb-d5ac212e11e7","Type":"ContainerDied","Data":"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c"} Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.313972 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de53ccf2-83b3-4116-a8eb-d5ac212e11e7","Type":"ContainerDied","Data":"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc"} Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.313994 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de53ccf2-83b3-4116-a8eb-d5ac212e11e7","Type":"ContainerDied","Data":"aaf03c72c4a7c24d8656ba98234f060e4ff324095c8d5dae0770c4fb6e763068"} Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.314015 4944 scope.go:117] "RemoveContainer" containerID="c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.316126 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.318555 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hpjlj" event={"ID":"6c338ca6-2795-4c94-b863-23579147b584","Type":"ContainerDied","Data":"8f5e3a19fc7d82256328b30ec36865a117cd5cf5680039e095ff5b7437e8ff6e"} Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.318692 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f5e3a19fc7d82256328b30ec36865a117cd5cf5680039e095ff5b7437e8ff6e" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.318847 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hpjlj" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.326635 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nst6k" event={"ID":"0493ffa7-e203-4f40-b117-07b07fe13380","Type":"ContainerDied","Data":"32c7d96b84faa5da891c5f38296c1230a00c462a301e85957b3c5e9ad07259f0"} Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.326842 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32c7d96b84faa5da891c5f38296c1230a00c462a301e85957b3c5e9ad07259f0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.326697 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nst6k" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.349288 4944 scope.go:117] "RemoveContainer" containerID="3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.366610 4944 scope.go:117] "RemoveContainer" containerID="4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.383270 4944 scope.go:117] "RemoveContainer" containerID="a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.399895 4944 scope.go:117] "RemoveContainer" containerID="c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a" Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.400268 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a\": container with ID starting with c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a not found: ID does not exist" containerID="c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.400299 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a"} err="failed to get container status \"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a\": rpc error: code = NotFound desc = could not find container \"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a\": container with ID starting with c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.400320 4944 scope.go:117] "RemoveContainer" containerID="3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5" Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.400506 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5\": container with ID starting with 3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5 not found: ID does not exist" containerID="3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.400528 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5"} err="failed to get container status \"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5\": rpc error: code = NotFound desc = could not find container \"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5\": container with ID starting with 3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5 not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.400541 4944 scope.go:117] "RemoveContainer" containerID="4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c" Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.400726 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c\": container with ID starting with 4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c not found: ID does not exist" containerID="4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.400746 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c"} err="failed to get container status \"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c\": rpc error: code = NotFound desc = could not find container \"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c\": container with ID starting with 4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.400757 4944 scope.go:117] "RemoveContainer" containerID="a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc" Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.400978 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc\": container with ID starting with a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc not found: ID does not exist" containerID="a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.400997 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc"} err="failed to get container status \"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc\": rpc error: code = NotFound desc = could not find container \"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc\": container with ID starting with a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.401011 4944 scope.go:117] "RemoveContainer" containerID="c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.401229 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a"} err="failed to get container status \"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a\": rpc error: code = NotFound desc = could not find container \"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a\": container with ID starting with c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.401246 4944 scope.go:117] "RemoveContainer" containerID="3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.401402 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5"} err="failed to get container status \"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5\": rpc error: code = NotFound desc = could not find container \"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5\": container with ID starting with 3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5 not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.401420 4944 scope.go:117] "RemoveContainer" containerID="4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.401581 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c"} err="failed to get container status \"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c\": rpc error: code = NotFound desc = could not find container \"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c\": container with ID starting with 4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.401597 4944 scope.go:117] "RemoveContainer" containerID="a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.401764 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc"} err="failed to get container status \"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc\": rpc error: code = NotFound desc = could not find container \"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc\": container with ID starting with a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.401779 4944 scope.go:117] "RemoveContainer" containerID="c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.401918 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a"} err="failed to get container status \"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a\": rpc error: code = NotFound desc = could not find container \"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a\": container with ID starting with c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.401936 4944 scope.go:117] "RemoveContainer" containerID="3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.402140 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5"} err="failed to get container status \"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5\": rpc error: code = NotFound desc = could not find container \"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5\": container with ID starting with 3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5 not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.402157 4944 scope.go:117] "RemoveContainer" containerID="4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.402330 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c"} err="failed to get container status \"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c\": rpc error: code = NotFound desc = could not find container \"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c\": container with ID starting with 4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.402347 4944 scope.go:117] "RemoveContainer" containerID="a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.402510 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc"} err="failed to get container status \"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc\": rpc error: code = NotFound desc = could not find container \"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc\": container with ID starting with a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.402525 4944 scope.go:117] "RemoveContainer" containerID="c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.402662 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a"} err="failed to get container status \"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a\": rpc error: code = NotFound desc = could not find container \"c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a\": container with ID starting with c2f5cc63dd129bd283d8dd94becf6e8f2fd09e2bdb2995ba50b134ed95886e2a not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.402677 4944 scope.go:117] "RemoveContainer" containerID="3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.402836 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5"} err="failed to get container status \"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5\": rpc error: code = NotFound desc = could not find container \"3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5\": container with ID starting with 3df416f567bd801e6124e2ed20042f924eb1832903b69ceae85b598de7aaf0a5 not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.402854 4944 scope.go:117] "RemoveContainer" containerID="4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.403009 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c"} err="failed to get container status \"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c\": rpc error: code = NotFound desc = could not find container \"4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c\": container with ID starting with 4a2d6d63bb62b41ad6c9a974f68040373e44cdd5527cbccfa1dd8ec4eed9924c not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.403024 4944 scope.go:117] "RemoveContainer" containerID="a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.403220 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc"} err="failed to get container status \"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc\": rpc error: code = NotFound desc = could not find container \"a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc\": container with ID starting with a0e077d2eeccec9b4feb57ebfb5eb47d7d01c02a3646d130f9dab3bffed33ddc not found: ID does not exist" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.425125 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-log-httpd\") pod \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.425165 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-config-data\") pod \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.425246 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wt9v\" (UniqueName: \"kubernetes.io/projected/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-kube-api-access-4wt9v\") pod \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.425302 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-scripts\") pod \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.425321 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-sg-core-conf-yaml\") pod \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.425396 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-combined-ca-bundle\") pod \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.425472 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-run-httpd\") pod \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\" (UID: \"de53ccf2-83b3-4116-a8eb-d5ac212e11e7\") " Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.426890 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "de53ccf2-83b3-4116-a8eb-d5ac212e11e7" (UID: "de53ccf2-83b3-4116-a8eb-d5ac212e11e7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.429722 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "de53ccf2-83b3-4116-a8eb-d5ac212e11e7" (UID: "de53ccf2-83b3-4116-a8eb-d5ac212e11e7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.430675 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-scripts" (OuterVolumeSpecName: "scripts") pod "de53ccf2-83b3-4116-a8eb-d5ac212e11e7" (UID: "de53ccf2-83b3-4116-a8eb-d5ac212e11e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.432543 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-kube-api-access-4wt9v" (OuterVolumeSpecName: "kube-api-access-4wt9v") pod "de53ccf2-83b3-4116-a8eb-d5ac212e11e7" (UID: "de53ccf2-83b3-4116-a8eb-d5ac212e11e7"). InnerVolumeSpecName "kube-api-access-4wt9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.454409 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "de53ccf2-83b3-4116-a8eb-d5ac212e11e7" (UID: "de53ccf2-83b3-4116-a8eb-d5ac212e11e7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.498283 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de53ccf2-83b3-4116-a8eb-d5ac212e11e7" (UID: "de53ccf2-83b3-4116-a8eb-d5ac212e11e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.528117 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.528157 4944 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.528170 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.528182 4944 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.528195 4944 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.528205 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wt9v\" (UniqueName: \"kubernetes.io/projected/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-kube-api-access-4wt9v\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.549941 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-config-data" (OuterVolumeSpecName: "config-data") pod "de53ccf2-83b3-4116-a8eb-d5ac212e11e7" (UID: "de53ccf2-83b3-4116-a8eb-d5ac212e11e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.603799 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.604264 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="proxy-httpd" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604283 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="proxy-httpd" Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.604297 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="ceilometer-notification-agent" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604304 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="ceilometer-notification-agent" Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.604320 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28e6918e-c11e-423e-996b-c43ee889155c" containerName="init" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604326 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="28e6918e-c11e-423e-996b-c43ee889155c" containerName="init" Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.604338 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="ceilometer-central-agent" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604344 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="ceilometer-central-agent" Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.604362 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="sg-core" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604368 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="sg-core" Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.604391 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c338ca6-2795-4c94-b863-23579147b584" containerName="cinder-db-sync" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604396 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c338ca6-2795-4c94-b863-23579147b584" containerName="cinder-db-sync" Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.604404 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28e6918e-c11e-423e-996b-c43ee889155c" containerName="dnsmasq-dns" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604411 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="28e6918e-c11e-423e-996b-c43ee889155c" containerName="dnsmasq-dns" Nov 24 09:11:20 crc kubenswrapper[4944]: E1124 09:11:20.604423 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0493ffa7-e203-4f40-b117-07b07fe13380" containerName="barbican-db-sync" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604429 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0493ffa7-e203-4f40-b117-07b07fe13380" containerName="barbican-db-sync" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604577 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="ceilometer-notification-agent" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604592 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="sg-core" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604603 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c338ca6-2795-4c94-b863-23579147b584" containerName="cinder-db-sync" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604610 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="proxy-httpd" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604624 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" containerName="ceilometer-central-agent" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604634 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="28e6918e-c11e-423e-996b-c43ee889155c" containerName="dnsmasq-dns" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.604642 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0493ffa7-e203-4f40-b117-07b07fe13380" containerName="barbican-db-sync" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.605589 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.610125 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.610410 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.622177 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-58f6cff88f-s86kk"] Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.624020 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.627556 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-58f6cff88f-s86kk"] Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.628458 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-j5rrf" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.629669 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de53ccf2-83b3-4116-a8eb-d5ac212e11e7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.631822 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qtvxl" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.632107 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.638125 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.639946 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.640146 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.644567 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7b47cdb5f8-g5n25"] Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.645950 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.656115 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.680936 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7b47cdb5f8-g5n25"] Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.732973 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733027 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733087 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ea29cb-ae71-48af-8005-ba99c178a573-logs\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733117 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h7gr\" (UniqueName: \"kubernetes.io/projected/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-kube-api-access-5h7gr\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733143 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-scripts\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733165 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733197 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733218 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733260 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dznxr\" (UniqueName: \"kubernetes.io/projected/e3ea29cb-ae71-48af-8005-ba99c178a573-kube-api-access-dznxr\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733291 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data-custom\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733341 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9rxv\" (UniqueName: \"kubernetes.io/projected/3d9c665f-2be7-4b08-934d-b63036656b5f-kube-api-access-p9rxv\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733389 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-combined-ca-bundle\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733421 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-combined-ca-bundle\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733445 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d9c665f-2be7-4b08-934d-b63036656b5f-logs\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733497 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.733544 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data-custom\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.754194 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-qxz6c"] Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.755707 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.829311 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-qxz6c"] Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.863185 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.866873 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data-custom\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.866942 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.866987 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867024 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867090 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ea29cb-ae71-48af-8005-ba99c178a573-logs\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867137 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h7gr\" (UniqueName: \"kubernetes.io/projected/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-kube-api-access-5h7gr\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867186 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-scripts\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867218 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867295 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867331 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867401 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dznxr\" (UniqueName: \"kubernetes.io/projected/e3ea29cb-ae71-48af-8005-ba99c178a573-kube-api-access-dznxr\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867441 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data-custom\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867534 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9rxv\" (UniqueName: \"kubernetes.io/projected/3d9c665f-2be7-4b08-934d-b63036656b5f-kube-api-access-p9rxv\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867575 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867607 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-config\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867634 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867704 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-combined-ca-bundle\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867756 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-combined-ca-bundle\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867780 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d9c665f-2be7-4b08-934d-b63036656b5f-logs\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867832 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-svc\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867880 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.867939 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f94gr\" (UniqueName: \"kubernetes.io/projected/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-kube-api-access-f94gr\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.879936 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d9c665f-2be7-4b08-934d-b63036656b5f-logs\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.885934 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ea29cb-ae71-48af-8005-ba99c178a573-logs\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.886014 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.929142 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-combined-ca-bundle\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.931030 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.955509 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.956009 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data-custom\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.961138 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.961386 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.962301 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data-custom\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.962827 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-combined-ca-bundle\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.963342 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.966185 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.972248 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.980807 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.980904 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.981019 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-config\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.981213 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-svc\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.981350 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f94gr\" (UniqueName: \"kubernetes.io/projected/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-kube-api-access-f94gr\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.981601 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.973270 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.982372 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.974449 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h7gr\" (UniqueName: \"kubernetes.io/projected/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-kube-api-access-5h7gr\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.982927 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-config\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.983035 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-svc\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.987672 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dznxr\" (UniqueName: \"kubernetes.io/projected/e3ea29cb-ae71-48af-8005-ba99c178a573-kube-api-access-dznxr\") pod \"barbican-keystone-listener-7b47cdb5f8-g5n25\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:20 crc kubenswrapper[4944]: I1124 09:11:20.988728 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.002040 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9rxv\" (UniqueName: \"kubernetes.io/projected/3d9c665f-2be7-4b08-934d-b63036656b5f-kube-api-access-p9rxv\") pod \"barbican-worker-58f6cff88f-s86kk\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.033196 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-scripts\") pod \"cinder-scheduler-0\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.077766 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f94gr\" (UniqueName: \"kubernetes.io/projected/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-kube-api-access-f94gr\") pod \"dnsmasq-dns-6578955fd5-qxz6c\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.089581 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.102343 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.105187 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.111934 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.130570 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.162151 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.200495 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.204534 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.204652 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-scripts\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.204740 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-log-httpd\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.204770 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.204865 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94vvf\" (UniqueName: \"kubernetes.io/projected/bb728e49-f81e-4f19-a087-67eda95ff876-kube-api-access-94vvf\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.204919 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-config-data\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.204955 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-run-httpd\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.214590 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.218949 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.223197 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.245203 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.275624 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.294370 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-58895d75f4-2z9z4"] Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.296081 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.308622 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.308789 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58895d75f4-2z9z4"] Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309148 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-log-httpd\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309203 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309236 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309284 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461d6a32-f9db-4344-8f96-737991c925ef-logs\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309330 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94vvf\" (UniqueName: \"kubernetes.io/projected/bb728e49-f81e-4f19-a087-67eda95ff876-kube-api-access-94vvf\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309363 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-config-data\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309398 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-run-httpd\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309435 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data-custom\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309461 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309528 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309584 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/461d6a32-f9db-4344-8f96-737991c925ef-etc-machine-id\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309617 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b6bn\" (UniqueName: \"kubernetes.io/projected/461d6a32-f9db-4344-8f96-737991c925ef-kube-api-access-8b6bn\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309706 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-scripts\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.309728 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-scripts\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.310488 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-log-httpd\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.316833 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-scripts\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.318187 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-config-data\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.320984 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-run-httpd\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.321091 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.322629 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.343911 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94vvf\" (UniqueName: \"kubernetes.io/projected/bb728e49-f81e-4f19-a087-67eda95ff876-kube-api-access-94vvf\") pod \"ceilometer-0\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.411492 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-scripts\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.411598 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdpf8\" (UniqueName: \"kubernetes.io/projected/3d6d9eee-7952-4b40-8958-2048e7cf7a58-kube-api-access-tdpf8\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.411670 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.411727 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461d6a32-f9db-4344-8f96-737991c925ef-logs\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.411750 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.411784 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d6d9eee-7952-4b40-8958-2048e7cf7a58-logs\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.411817 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data-custom\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.411897 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-combined-ca-bundle\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.411982 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data-custom\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.412014 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.412041 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/461d6a32-f9db-4344-8f96-737991c925ef-etc-machine-id\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.412092 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b6bn\" (UniqueName: \"kubernetes.io/projected/461d6a32-f9db-4344-8f96-737991c925ef-kube-api-access-8b6bn\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.415467 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/461d6a32-f9db-4344-8f96-737991c925ef-etc-machine-id\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.415780 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461d6a32-f9db-4344-8f96-737991c925ef-logs\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.419533 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.420657 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-scripts\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.425730 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data-custom\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.425754 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.431699 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b6bn\" (UniqueName: \"kubernetes.io/projected/461d6a32-f9db-4344-8f96-737991c925ef-kube-api-access-8b6bn\") pod \"cinder-api-0\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.475960 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.514272 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdpf8\" (UniqueName: \"kubernetes.io/projected/3d6d9eee-7952-4b40-8958-2048e7cf7a58-kube-api-access-tdpf8\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.514445 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.514495 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d6d9eee-7952-4b40-8958-2048e7cf7a58-logs\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.514523 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data-custom\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.514593 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-combined-ca-bundle\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.515444 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d6d9eee-7952-4b40-8958-2048e7cf7a58-logs\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.519979 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data-custom\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.520704 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-combined-ca-bundle\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.528340 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.530794 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdpf8\" (UniqueName: \"kubernetes.io/projected/3d6d9eee-7952-4b40-8958-2048e7cf7a58-kube-api-access-tdpf8\") pod \"barbican-api-58895d75f4-2z9z4\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.544803 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.635512 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.714405 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7b47cdb5f8-g5n25"] Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.829099 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-qxz6c"] Nov 24 09:11:21 crc kubenswrapper[4944]: I1124 09:11:21.978298 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.003413 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-58f6cff88f-s86kk"] Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.092447 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.210803 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.235879 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58895d75f4-2z9z4"] Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.295752 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de53ccf2-83b3-4116-a8eb-d5ac212e11e7" path="/var/lib/kubelet/pods/de53ccf2-83b3-4116-a8eb-d5ac212e11e7/volumes" Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.375020 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" event={"ID":"e3ea29cb-ae71-48af-8005-ba99c178a573","Type":"ContainerStarted","Data":"9ff6376fa9603b040179576baa880c926be1eeeff20bc399721dcd84b1c41fe8"} Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.381671 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-58f6cff88f-s86kk" event={"ID":"3d9c665f-2be7-4b08-934d-b63036656b5f","Type":"ContainerStarted","Data":"eb0ffeae65b43b2ec46ada236b5d55d28eb59b258a8954c9cfafb8df9eb31e28"} Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.387849 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f","Type":"ContainerStarted","Data":"ebb8a256d39a3cffbada4c1392f70f04b3f84af91970dff5cff6914e0807e489"} Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.392478 4944 generic.go:334] "Generic (PLEG): container finished" podID="9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" containerID="33286e6f5da914e557f15218ff1cd6235479db1db7c60eb8b559b8d60d6343b5" exitCode=0 Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.392537 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" event={"ID":"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50","Type":"ContainerDied","Data":"33286e6f5da914e557f15218ff1cd6235479db1db7c60eb8b559b8d60d6343b5"} Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.392562 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" event={"ID":"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50","Type":"ContainerStarted","Data":"7e24ee81704d1fd5d02b8d2a46731a59de09fa79b74c108799ccf0e29f71703c"} Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.398495 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58895d75f4-2z9z4" event={"ID":"3d6d9eee-7952-4b40-8958-2048e7cf7a58","Type":"ContainerStarted","Data":"9535ba1dfc6a070b42edc00506fa447d01cdf369bca76d5a87712c431887bab7"} Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.400557 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"461d6a32-f9db-4344-8f96-737991c925ef","Type":"ContainerStarted","Data":"e0b2f08216da5d45a7c35682fffb42705bc30f121cd53fa57337453f97c7cb60"} Nov 24 09:11:22 crc kubenswrapper[4944]: I1124 09:11:22.401478 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb728e49-f81e-4f19-a087-67eda95ff876","Type":"ContainerStarted","Data":"58f421e1b79935b6222f2f2f2be71c3f97552ae618693c3307af05a7e39d0a6d"} Nov 24 09:11:23 crc kubenswrapper[4944]: I1124 09:11:23.074704 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:11:23 crc kubenswrapper[4944]: I1124 09:11:23.418374 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" event={"ID":"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50","Type":"ContainerStarted","Data":"bf0729a1ce61e434d095faba09fa37d27712f335d1509123cdcf54575a7ac2a4"} Nov 24 09:11:23 crc kubenswrapper[4944]: I1124 09:11:23.420952 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:23 crc kubenswrapper[4944]: I1124 09:11:23.423269 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58895d75f4-2z9z4" event={"ID":"3d6d9eee-7952-4b40-8958-2048e7cf7a58","Type":"ContainerStarted","Data":"45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37"} Nov 24 09:11:23 crc kubenswrapper[4944]: I1124 09:11:23.425045 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"461d6a32-f9db-4344-8f96-737991c925ef","Type":"ContainerStarted","Data":"07a7c5d14487184e00be4545608a72603f1b3f56553d912070d2230c3f22b18c"} Nov 24 09:11:23 crc kubenswrapper[4944]: I1124 09:11:23.445730 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" podStartSLOduration=3.445714165 podStartE2EDuration="3.445714165s" podCreationTimestamp="2025-11-24 09:11:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:23.440013483 +0000 UTC m=+1143.974453955" watchObservedRunningTime="2025-11-24 09:11:23.445714165 +0000 UTC m=+1143.980154627" Nov 24 09:11:23 crc kubenswrapper[4944]: I1124 09:11:23.548854 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:11:23 crc kubenswrapper[4944]: I1124 09:11:23.549154 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:11:23 crc kubenswrapper[4944]: I1124 09:11:23.549253 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:11:23 crc kubenswrapper[4944]: I1124 09:11:23.550390 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d4676a5ed72cda8e3ecd73f73356ca831b52b82f176ed2715695ec66c7c950e"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:11:23 crc kubenswrapper[4944]: I1124 09:11:23.550518 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://1d4676a5ed72cda8e3ecd73f73356ca831b52b82f176ed2715695ec66c7c950e" gracePeriod=600 Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.439544 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb728e49-f81e-4f19-a087-67eda95ff876","Type":"ContainerStarted","Data":"5b9e9df6d119e7e9ed26178412db428957957754d818ba3fd5e628bfb2e730a4"} Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.444017 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="1d4676a5ed72cda8e3ecd73f73356ca831b52b82f176ed2715695ec66c7c950e" exitCode=0 Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.444108 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"1d4676a5ed72cda8e3ecd73f73356ca831b52b82f176ed2715695ec66c7c950e"} Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.444139 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"f8ac17962023518b7abc2c1aa6e647faf7d9a5693323d499415d0670598b6a65"} Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.444163 4944 scope.go:117] "RemoveContainer" containerID="b51cb7736616a4e1de4d8a3f423bb2e041c1d8aa22c653e62b3b1a70f4ce6afd" Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.452224 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" event={"ID":"e3ea29cb-ae71-48af-8005-ba99c178a573","Type":"ContainerStarted","Data":"e3db2c95e6254ee19d5cd78241174e4c4f6100bc9626f271af3b94acb76058e6"} Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.452275 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" event={"ID":"e3ea29cb-ae71-48af-8005-ba99c178a573","Type":"ContainerStarted","Data":"5e2bf771bd1be4b28e3fa735cec21fcbbccd1e4807f6ac424dd4405dc8f8a329"} Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.459872 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-58f6cff88f-s86kk" event={"ID":"3d9c665f-2be7-4b08-934d-b63036656b5f","Type":"ContainerStarted","Data":"9906021955dc6929c0ac49f16b105afc15f33b120aab600fc31ecf2026e5ba6e"} Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.459918 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-58f6cff88f-s86kk" event={"ID":"3d9c665f-2be7-4b08-934d-b63036656b5f","Type":"ContainerStarted","Data":"219c63bcbbda1c8ee292c94dcc9e2f39b33c5aab0d972baaee0d1dc1f454692d"} Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.462698 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f","Type":"ContainerStarted","Data":"9b26a48292d2ccb1f795f8abcb1e2ff6c90883cd75fd2d2912844af6ac07e0a3"} Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.477360 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58895d75f4-2z9z4" event={"ID":"3d6d9eee-7952-4b40-8958-2048e7cf7a58","Type":"ContainerStarted","Data":"a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b"} Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.478886 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.478938 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.495638 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" podStartSLOduration=2.7799833769999998 podStartE2EDuration="4.495611196s" podCreationTimestamp="2025-11-24 09:11:20 +0000 UTC" firstStartedPulling="2025-11-24 09:11:21.771872737 +0000 UTC m=+1142.306313199" lastFinishedPulling="2025-11-24 09:11:23.487500556 +0000 UTC m=+1144.021941018" observedRunningTime="2025-11-24 09:11:24.483313214 +0000 UTC m=+1145.017753676" watchObservedRunningTime="2025-11-24 09:11:24.495611196 +0000 UTC m=+1145.030051668" Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.497101 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="461d6a32-f9db-4344-8f96-737991c925ef" containerName="cinder-api-log" containerID="cri-o://07a7c5d14487184e00be4545608a72603f1b3f56553d912070d2230c3f22b18c" gracePeriod=30 Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.497359 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="461d6a32-f9db-4344-8f96-737991c925ef" containerName="cinder-api" containerID="cri-o://18258110a4c5ed8692ca22a715d11dcc1f4e93846569f14ad958bb3c2936dec4" gracePeriod=30 Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.497965 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"461d6a32-f9db-4344-8f96-737991c925ef","Type":"ContainerStarted","Data":"18258110a4c5ed8692ca22a715d11dcc1f4e93846569f14ad958bb3c2936dec4"} Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.498157 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.524565 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-58f6cff88f-s86kk" podStartSLOduration=3.043736242 podStartE2EDuration="4.524547748s" podCreationTimestamp="2025-11-24 09:11:20 +0000 UTC" firstStartedPulling="2025-11-24 09:11:22.007755814 +0000 UTC m=+1142.542196276" lastFinishedPulling="2025-11-24 09:11:23.48856732 +0000 UTC m=+1144.023007782" observedRunningTime="2025-11-24 09:11:24.514041243 +0000 UTC m=+1145.048481715" watchObservedRunningTime="2025-11-24 09:11:24.524547748 +0000 UTC m=+1145.058988200" Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.549317 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.5492856459999995 podStartE2EDuration="4.549285646s" podCreationTimestamp="2025-11-24 09:11:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:24.539865367 +0000 UTC m=+1145.074305829" watchObservedRunningTime="2025-11-24 09:11:24.549285646 +0000 UTC m=+1145.083726108" Nov 24 09:11:24 crc kubenswrapper[4944]: I1124 09:11:24.585502 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-58895d75f4-2z9z4" podStartSLOduration=3.585485211 podStartE2EDuration="3.585485211s" podCreationTimestamp="2025-11-24 09:11:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:24.564662227 +0000 UTC m=+1145.099102709" watchObservedRunningTime="2025-11-24 09:11:24.585485211 +0000 UTC m=+1145.119925673" Nov 24 09:11:25 crc kubenswrapper[4944]: I1124 09:11:25.509580 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb728e49-f81e-4f19-a087-67eda95ff876","Type":"ContainerStarted","Data":"695684c6c6e8fb9277e670253bd334e3156a4594ad89f89cffb548da35d8d40b"} Nov 24 09:11:25 crc kubenswrapper[4944]: I1124 09:11:25.510259 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb728e49-f81e-4f19-a087-67eda95ff876","Type":"ContainerStarted","Data":"e848fb6c2ab71eb4d88e062d7e414fe1e507449e413fc03eb99ca44255bc3f33"} Nov 24 09:11:25 crc kubenswrapper[4944]: I1124 09:11:25.517811 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f","Type":"ContainerStarted","Data":"fb3a566164b01b8e8fd3ba7b81d506f9d0534db28b9dba12132e4ec72253f454"} Nov 24 09:11:25 crc kubenswrapper[4944]: I1124 09:11:25.539878 4944 generic.go:334] "Generic (PLEG): container finished" podID="461d6a32-f9db-4344-8f96-737991c925ef" containerID="07a7c5d14487184e00be4545608a72603f1b3f56553d912070d2230c3f22b18c" exitCode=143 Nov 24 09:11:25 crc kubenswrapper[4944]: I1124 09:11:25.541341 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"461d6a32-f9db-4344-8f96-737991c925ef","Type":"ContainerDied","Data":"07a7c5d14487184e00be4545608a72603f1b3f56553d912070d2230c3f22b18c"} Nov 24 09:11:25 crc kubenswrapper[4944]: I1124 09:11:25.565726 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.076333333 podStartE2EDuration="5.565704861s" podCreationTimestamp="2025-11-24 09:11:20 +0000 UTC" firstStartedPulling="2025-11-24 09:11:21.997662883 +0000 UTC m=+1142.532103345" lastFinishedPulling="2025-11-24 09:11:23.487034411 +0000 UTC m=+1144.021474873" observedRunningTime="2025-11-24 09:11:25.554619738 +0000 UTC m=+1146.089060210" watchObservedRunningTime="2025-11-24 09:11:25.565704861 +0000 UTC m=+1146.100145323" Nov 24 09:11:26 crc kubenswrapper[4944]: I1124 09:11:26.224328 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.207297 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5f49db88bb-dzrb6"] Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.209555 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.211863 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.220809 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.225383 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-combined-ca-bundle\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.225523 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-public-tls-certs\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.225582 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data-custom\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.225615 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-internal-tls-certs\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.225887 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzczg\" (UniqueName: \"kubernetes.io/projected/046a8c15-6acd-47fe-aef7-daac5085c2ec-kube-api-access-dzczg\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.226076 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/046a8c15-6acd-47fe-aef7-daac5085c2ec-logs\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.226376 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.228899 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f49db88bb-dzrb6"] Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.329696 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-combined-ca-bundle\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.329796 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-public-tls-certs\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.329838 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data-custom\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.329871 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-internal-tls-certs\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.329951 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzczg\" (UniqueName: \"kubernetes.io/projected/046a8c15-6acd-47fe-aef7-daac5085c2ec-kube-api-access-dzczg\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.329998 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/046a8c15-6acd-47fe-aef7-daac5085c2ec-logs\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.330097 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.332092 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/046a8c15-6acd-47fe-aef7-daac5085c2ec-logs\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.340603 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.340800 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-public-tls-certs\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.344636 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-combined-ca-bundle\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.348675 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data-custom\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.348848 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzczg\" (UniqueName: \"kubernetes.io/projected/046a8c15-6acd-47fe-aef7-daac5085c2ec-kube-api-access-dzczg\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.364530 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-internal-tls-certs\") pod \"barbican-api-5f49db88bb-dzrb6\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.527994 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.561443 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb728e49-f81e-4f19-a087-67eda95ff876","Type":"ContainerStarted","Data":"b322645c4ac77486a9ce3db6ba3bd6419053bba567bd17251759b44dbb6f547e"} Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.561524 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.901601 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.368691149 podStartE2EDuration="7.901575679s" podCreationTimestamp="2025-11-24 09:11:20 +0000 UTC" firstStartedPulling="2025-11-24 09:11:22.103415353 +0000 UTC m=+1142.637855815" lastFinishedPulling="2025-11-24 09:11:26.636299883 +0000 UTC m=+1147.170740345" observedRunningTime="2025-11-24 09:11:27.586346783 +0000 UTC m=+1148.120787265" watchObservedRunningTime="2025-11-24 09:11:27.901575679 +0000 UTC m=+1148.436016151" Nov 24 09:11:27 crc kubenswrapper[4944]: I1124 09:11:27.905780 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f49db88bb-dzrb6"] Nov 24 09:11:28 crc kubenswrapper[4944]: I1124 09:11:28.569877 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f49db88bb-dzrb6" event={"ID":"046a8c15-6acd-47fe-aef7-daac5085c2ec","Type":"ContainerStarted","Data":"01f911e69ae02d9b8cfae324442ea1d08c8cc067ff998c1912785fe2fe020c3c"} Nov 24 09:11:29 crc kubenswrapper[4944]: I1124 09:11:29.588890 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f49db88bb-dzrb6" event={"ID":"046a8c15-6acd-47fe-aef7-daac5085c2ec","Type":"ContainerStarted","Data":"7aa8ffc64bce11fa46956c14a0b6ed75513f2964c1dd926a5e6be5120e48e88e"} Nov 24 09:11:29 crc kubenswrapper[4944]: I1124 09:11:29.937569 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:11:30 crc kubenswrapper[4944]: I1124 09:11:30.598320 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f49db88bb-dzrb6" event={"ID":"046a8c15-6acd-47fe-aef7-daac5085c2ec","Type":"ContainerStarted","Data":"7159a46717426d269e3e9ca766c99432598e2da70a9cf82e6a43abda53ec27b9"} Nov 24 09:11:30 crc kubenswrapper[4944]: I1124 09:11:30.599268 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:30 crc kubenswrapper[4944]: I1124 09:11:30.599341 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:30 crc kubenswrapper[4944]: I1124 09:11:30.619864 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5f49db88bb-dzrb6" podStartSLOduration=3.619134211 podStartE2EDuration="3.619134211s" podCreationTimestamp="2025-11-24 09:11:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:30.613170341 +0000 UTC m=+1151.147610803" watchObservedRunningTime="2025-11-24 09:11:30.619134211 +0000 UTC m=+1151.153574663" Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.091875 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.186270 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-795tt"] Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.187362 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-795tt" podUID="7b35c8ac-158b-480d-9dcc-d0a474151efa" containerName="dnsmasq-dns" containerID="cri-o://b9dc3d12c1db828a9cd9ce34593a1437825bdb491b2df896b66efac37ce2a267" gracePeriod=10 Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.559892 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.646761 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.648740 4944 generic.go:334] "Generic (PLEG): container finished" podID="7b35c8ac-158b-480d-9dcc-d0a474151efa" containerID="b9dc3d12c1db828a9cd9ce34593a1437825bdb491b2df896b66efac37ce2a267" exitCode=0 Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.649047 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" containerName="cinder-scheduler" containerID="cri-o://9b26a48292d2ccb1f795f8abcb1e2ff6c90883cd75fd2d2912844af6ac07e0a3" gracePeriod=30 Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.649520 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-795tt" event={"ID":"7b35c8ac-158b-480d-9dcc-d0a474151efa","Type":"ContainerDied","Data":"b9dc3d12c1db828a9cd9ce34593a1437825bdb491b2df896b66efac37ce2a267"} Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.651129 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" containerName="probe" containerID="cri-o://fb3a566164b01b8e8fd3ba7b81d506f9d0534db28b9dba12132e4ec72253f454" gracePeriod=30 Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.780371 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.942697 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-config\") pod \"7b35c8ac-158b-480d-9dcc-d0a474151efa\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.942748 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8msfp\" (UniqueName: \"kubernetes.io/projected/7b35c8ac-158b-480d-9dcc-d0a474151efa-kube-api-access-8msfp\") pod \"7b35c8ac-158b-480d-9dcc-d0a474151efa\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.942822 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-nb\") pod \"7b35c8ac-158b-480d-9dcc-d0a474151efa\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.942953 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-sb\") pod \"7b35c8ac-158b-480d-9dcc-d0a474151efa\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.943008 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-swift-storage-0\") pod \"7b35c8ac-158b-480d-9dcc-d0a474151efa\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.943216 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-svc\") pod \"7b35c8ac-158b-480d-9dcc-d0a474151efa\" (UID: \"7b35c8ac-158b-480d-9dcc-d0a474151efa\") " Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.963940 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b35c8ac-158b-480d-9dcc-d0a474151efa-kube-api-access-8msfp" (OuterVolumeSpecName: "kube-api-access-8msfp") pod "7b35c8ac-158b-480d-9dcc-d0a474151efa" (UID: "7b35c8ac-158b-480d-9dcc-d0a474151efa"). InnerVolumeSpecName "kube-api-access-8msfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:31 crc kubenswrapper[4944]: I1124 09:11:31.993838 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7b35c8ac-158b-480d-9dcc-d0a474151efa" (UID: "7b35c8ac-158b-480d-9dcc-d0a474151efa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.005755 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7b35c8ac-158b-480d-9dcc-d0a474151efa" (UID: "7b35c8ac-158b-480d-9dcc-d0a474151efa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.009409 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-config" (OuterVolumeSpecName: "config") pod "7b35c8ac-158b-480d-9dcc-d0a474151efa" (UID: "7b35c8ac-158b-480d-9dcc-d0a474151efa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.010008 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7b35c8ac-158b-480d-9dcc-d0a474151efa" (UID: "7b35c8ac-158b-480d-9dcc-d0a474151efa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.024188 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7b35c8ac-158b-480d-9dcc-d0a474151efa" (UID: "7b35c8ac-158b-480d-9dcc-d0a474151efa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.045737 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.045779 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.045796 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8msfp\" (UniqueName: \"kubernetes.io/projected/7b35c8ac-158b-480d-9dcc-d0a474151efa-kube-api-access-8msfp\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.045813 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.045825 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.045839 4944 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b35c8ac-158b-480d-9dcc-d0a474151efa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.658982 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-795tt" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.658966 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-795tt" event={"ID":"7b35c8ac-158b-480d-9dcc-d0a474151efa","Type":"ContainerDied","Data":"720a0bf11e716b8b0e59df8f177d4d811679f30ad0cb3124b26ab35aef17d04c"} Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.660199 4944 scope.go:117] "RemoveContainer" containerID="b9dc3d12c1db828a9cd9ce34593a1437825bdb491b2df896b66efac37ce2a267" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.663024 4944 generic.go:334] "Generic (PLEG): container finished" podID="34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" containerID="fb3a566164b01b8e8fd3ba7b81d506f9d0534db28b9dba12132e4ec72253f454" exitCode=0 Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.663089 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f","Type":"ContainerDied","Data":"fb3a566164b01b8e8fd3ba7b81d506f9d0534db28b9dba12132e4ec72253f454"} Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.688779 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-795tt"] Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.696434 4944 scope.go:117] "RemoveContainer" containerID="7bbaef1f0abf82ee381df9b478d9430d56e85b59230e0d8ecb5f76b4d607d961" Nov 24 09:11:32 crc kubenswrapper[4944]: I1124 09:11:32.698457 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-795tt"] Nov 24 09:11:33 crc kubenswrapper[4944]: I1124 09:11:33.314060 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:33 crc kubenswrapper[4944]: I1124 09:11:33.389348 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:33 crc kubenswrapper[4944]: I1124 09:11:33.714942 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 09:11:34 crc kubenswrapper[4944]: I1124 09:11:34.044390 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:11:34 crc kubenswrapper[4944]: I1124 09:11:34.109572 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-54f69f9546-4ksss"] Nov 24 09:11:34 crc kubenswrapper[4944]: I1124 09:11:34.109846 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-54f69f9546-4ksss" podUID="0c922b8a-f6f5-4801-839c-e67fcef0a67b" containerName="neutron-api" containerID="cri-o://02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec" gracePeriod=30 Nov 24 09:11:34 crc kubenswrapper[4944]: I1124 09:11:34.109895 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-54f69f9546-4ksss" podUID="0c922b8a-f6f5-4801-839c-e67fcef0a67b" containerName="neutron-httpd" containerID="cri-o://82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a" gracePeriod=30 Nov 24 09:11:34 crc kubenswrapper[4944]: I1124 09:11:34.291564 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b35c8ac-158b-480d-9dcc-d0a474151efa" path="/var/lib/kubelet/pods/7b35c8ac-158b-480d-9dcc-d0a474151efa/volumes" Nov 24 09:11:34 crc kubenswrapper[4944]: I1124 09:11:34.596169 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:11:34 crc kubenswrapper[4944]: I1124 09:11:34.716470 4944 generic.go:334] "Generic (PLEG): container finished" podID="0c922b8a-f6f5-4801-839c-e67fcef0a67b" containerID="82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a" exitCode=0 Nov 24 09:11:34 crc kubenswrapper[4944]: I1124 09:11:34.716518 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f69f9546-4ksss" event={"ID":"0c922b8a-f6f5-4801-839c-e67fcef0a67b","Type":"ContainerDied","Data":"82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a"} Nov 24 09:11:35 crc kubenswrapper[4944]: I1124 09:11:35.731716 4944 generic.go:334] "Generic (PLEG): container finished" podID="34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" containerID="9b26a48292d2ccb1f795f8abcb1e2ff6c90883cd75fd2d2912844af6ac07e0a3" exitCode=0 Nov 24 09:11:35 crc kubenswrapper[4944]: I1124 09:11:35.732039 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f","Type":"ContainerDied","Data":"9b26a48292d2ccb1f795f8abcb1e2ff6c90883cd75fd2d2912844af6ac07e0a3"} Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.066211 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.228993 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-combined-ca-bundle\") pod \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.229064 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h7gr\" (UniqueName: \"kubernetes.io/projected/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-kube-api-access-5h7gr\") pod \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.229136 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data-custom\") pod \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.229180 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-etc-machine-id\") pod \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.229246 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-scripts\") pod \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.229281 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data\") pod \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\" (UID: \"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f\") " Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.230005 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" (UID: "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.235859 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" (UID: "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.236178 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-kube-api-access-5h7gr" (OuterVolumeSpecName: "kube-api-access-5h7gr") pod "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" (UID: "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f"). InnerVolumeSpecName "kube-api-access-5h7gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.238288 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-scripts" (OuterVolumeSpecName: "scripts") pod "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" (UID: "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.280644 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" (UID: "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.328956 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data" (OuterVolumeSpecName: "config-data") pod "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" (UID: "34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.331061 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.331343 4944 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.331361 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.331369 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.331378 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.331386 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h7gr\" (UniqueName: \"kubernetes.io/projected/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f-kube-api-access-5h7gr\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.741809 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f","Type":"ContainerDied","Data":"ebb8a256d39a3cffbada4c1392f70f04b3f84af91970dff5cff6914e0807e489"} Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.742176 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.742179 4944 scope.go:117] "RemoveContainer" containerID="fb3a566164b01b8e8fd3ba7b81d506f9d0534db28b9dba12132e4ec72253f454" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.769176 4944 scope.go:117] "RemoveContainer" containerID="9b26a48292d2ccb1f795f8abcb1e2ff6c90883cd75fd2d2912844af6ac07e0a3" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.773165 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.784690 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.807815 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:11:36 crc kubenswrapper[4944]: E1124 09:11:36.808610 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b35c8ac-158b-480d-9dcc-d0a474151efa" containerName="init" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.808634 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b35c8ac-158b-480d-9dcc-d0a474151efa" containerName="init" Nov 24 09:11:36 crc kubenswrapper[4944]: E1124 09:11:36.808651 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b35c8ac-158b-480d-9dcc-d0a474151efa" containerName="dnsmasq-dns" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.808659 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b35c8ac-158b-480d-9dcc-d0a474151efa" containerName="dnsmasq-dns" Nov 24 09:11:36 crc kubenswrapper[4944]: E1124 09:11:36.808685 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" containerName="probe" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.808694 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" containerName="probe" Nov 24 09:11:36 crc kubenswrapper[4944]: E1124 09:11:36.808716 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" containerName="cinder-scheduler" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.808726 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" containerName="cinder-scheduler" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.808953 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" containerName="cinder-scheduler" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.808977 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b35c8ac-158b-480d-9dcc-d0a474151efa" containerName="dnsmasq-dns" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.809000 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" containerName="probe" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.811720 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.821665 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.821681 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.941291 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-scripts\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.941380 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53d02d15-b171-4e78-b528-9bec6362bf70-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.941520 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.941720 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.941804 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:36 crc kubenswrapper[4944]: I1124 09:11:36.942034 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk6fn\" (UniqueName: \"kubernetes.io/projected/53d02d15-b171-4e78-b528-9bec6362bf70-kube-api-access-nk6fn\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.043872 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk6fn\" (UniqueName: \"kubernetes.io/projected/53d02d15-b171-4e78-b528-9bec6362bf70-kube-api-access-nk6fn\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.044013 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-scripts\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.044097 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53d02d15-b171-4e78-b528-9bec6362bf70-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.044134 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.044181 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.044206 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.044227 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53d02d15-b171-4e78-b528-9bec6362bf70-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.052169 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-scripts\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.052586 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.054333 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.064771 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.076967 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk6fn\" (UniqueName: \"kubernetes.io/projected/53d02d15-b171-4e78-b528-9bec6362bf70-kube-api-access-nk6fn\") pod \"cinder-scheduler-0\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.135712 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.441676 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:11:37 crc kubenswrapper[4944]: W1124 09:11:37.452316 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53d02d15_b171_4e78_b528_9bec6362bf70.slice/crio-8a40812f7cb63d0cb175cee00130c9703f7a8c995b196814512aaf1b5b9db410 WatchSource:0}: Error finding container 8a40812f7cb63d0cb175cee00130c9703f7a8c995b196814512aaf1b5b9db410: Status 404 returned error can't find the container with id 8a40812f7cb63d0cb175cee00130c9703f7a8c995b196814512aaf1b5b9db410 Nov 24 09:11:37 crc kubenswrapper[4944]: I1124 09:11:37.753505 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"53d02d15-b171-4e78-b528-9bec6362bf70","Type":"ContainerStarted","Data":"8a40812f7cb63d0cb175cee00130c9703f7a8c995b196814512aaf1b5b9db410"} Nov 24 09:11:38 crc kubenswrapper[4944]: I1124 09:11:38.293626 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f" path="/var/lib/kubelet/pods/34c3682d-2b86-4be1-b9e2-e3e1f2c55e8f/volumes" Nov 24 09:11:38 crc kubenswrapper[4944]: I1124 09:11:38.776766 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"53d02d15-b171-4e78-b528-9bec6362bf70","Type":"ContainerStarted","Data":"b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b"} Nov 24 09:11:38 crc kubenswrapper[4944]: I1124 09:11:38.910216 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:11:39 crc kubenswrapper[4944]: I1124 09:11:39.194321 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:39 crc kubenswrapper[4944]: I1124 09:11:39.609610 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:11:39 crc kubenswrapper[4944]: I1124 09:11:39.688190 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-58895d75f4-2z9z4"] Nov 24 09:11:39 crc kubenswrapper[4944]: I1124 09:11:39.688817 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-58895d75f4-2z9z4" podUID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerName="barbican-api" containerID="cri-o://a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b" gracePeriod=30 Nov 24 09:11:39 crc kubenswrapper[4944]: I1124 09:11:39.688468 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-58895d75f4-2z9z4" podUID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerName="barbican-api-log" containerID="cri-o://45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37" gracePeriod=30 Nov 24 09:11:39 crc kubenswrapper[4944]: I1124 09:11:39.830620 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"53d02d15-b171-4e78-b528-9bec6362bf70","Type":"ContainerStarted","Data":"fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f"} Nov 24 09:11:39 crc kubenswrapper[4944]: I1124 09:11:39.874425 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.874408561 podStartE2EDuration="3.874408561s" podCreationTimestamp="2025-11-24 09:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:39.869540566 +0000 UTC m=+1160.403981028" watchObservedRunningTime="2025-11-24 09:11:39.874408561 +0000 UTC m=+1160.408849023" Nov 24 09:11:40 crc kubenswrapper[4944]: I1124 09:11:40.842917 4944 generic.go:334] "Generic (PLEG): container finished" podID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerID="45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37" exitCode=143 Nov 24 09:11:40 crc kubenswrapper[4944]: I1124 09:11:40.843885 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58895d75f4-2z9z4" event={"ID":"3d6d9eee-7952-4b40-8958-2048e7cf7a58","Type":"ContainerDied","Data":"45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37"} Nov 24 09:11:42 crc kubenswrapper[4944]: I1124 09:11:42.136459 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 09:11:42 crc kubenswrapper[4944]: I1124 09:11:42.877524 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-58895d75f4-2z9z4" podUID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:53814->10.217.0.160:9311: read: connection reset by peer" Nov 24 09:11:42 crc kubenswrapper[4944]: I1124 09:11:42.877532 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-58895d75f4-2z9z4" podUID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:53830->10.217.0.160:9311: read: connection reset by peer" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.307622 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.474415 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdpf8\" (UniqueName: \"kubernetes.io/projected/3d6d9eee-7952-4b40-8958-2048e7cf7a58-kube-api-access-tdpf8\") pod \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.474497 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-combined-ca-bundle\") pod \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.474535 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data-custom\") pod \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.474564 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d6d9eee-7952-4b40-8958-2048e7cf7a58-logs\") pod \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.474656 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data\") pod \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\" (UID: \"3d6d9eee-7952-4b40-8958-2048e7cf7a58\") " Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.475150 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d6d9eee-7952-4b40-8958-2048e7cf7a58-logs" (OuterVolumeSpecName: "logs") pod "3d6d9eee-7952-4b40-8958-2048e7cf7a58" (UID: "3d6d9eee-7952-4b40-8958-2048e7cf7a58"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.476070 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d6d9eee-7952-4b40-8958-2048e7cf7a58-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.480284 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3d6d9eee-7952-4b40-8958-2048e7cf7a58" (UID: "3d6d9eee-7952-4b40-8958-2048e7cf7a58"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.496562 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d6d9eee-7952-4b40-8958-2048e7cf7a58-kube-api-access-tdpf8" (OuterVolumeSpecName: "kube-api-access-tdpf8") pod "3d6d9eee-7952-4b40-8958-2048e7cf7a58" (UID: "3d6d9eee-7952-4b40-8958-2048e7cf7a58"). InnerVolumeSpecName "kube-api-access-tdpf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.506251 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d6d9eee-7952-4b40-8958-2048e7cf7a58" (UID: "3d6d9eee-7952-4b40-8958-2048e7cf7a58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.537235 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data" (OuterVolumeSpecName: "config-data") pod "3d6d9eee-7952-4b40-8958-2048e7cf7a58" (UID: "3d6d9eee-7952-4b40-8958-2048e7cf7a58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.577849 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdpf8\" (UniqueName: \"kubernetes.io/projected/3d6d9eee-7952-4b40-8958-2048e7cf7a58-kube-api-access-tdpf8\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.578143 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.578217 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.578285 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d6d9eee-7952-4b40-8958-2048e7cf7a58-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.761452 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 09:11:43 crc kubenswrapper[4944]: E1124 09:11:43.761834 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerName="barbican-api" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.761851 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerName="barbican-api" Nov 24 09:11:43 crc kubenswrapper[4944]: E1124 09:11:43.761880 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerName="barbican-api-log" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.761886 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerName="barbican-api-log" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.762097 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerName="barbican-api-log" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.762116 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerName="barbican-api" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.762675 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.765164 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-8b9xp" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.766015 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.766357 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.782479 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.871391 4944 generic.go:334] "Generic (PLEG): container finished" podID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" containerID="a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b" exitCode=0 Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.871520 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58895d75f4-2z9z4" event={"ID":"3d6d9eee-7952-4b40-8958-2048e7cf7a58","Type":"ContainerDied","Data":"a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b"} Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.871522 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58895d75f4-2z9z4" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.871756 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58895d75f4-2z9z4" event={"ID":"3d6d9eee-7952-4b40-8958-2048e7cf7a58","Type":"ContainerDied","Data":"9535ba1dfc6a070b42edc00506fa447d01cdf369bca76d5a87712c431887bab7"} Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.871801 4944 scope.go:117] "RemoveContainer" containerID="a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.883157 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.883253 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.883336 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config-secret\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.883368 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw9rf\" (UniqueName: \"kubernetes.io/projected/9a8dab89-ec08-4497-a890-bcb70c8ea430-kube-api-access-bw9rf\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.896033 4944 scope.go:117] "RemoveContainer" containerID="45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.912884 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-58895d75f4-2z9z4"] Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.921879 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-58895d75f4-2z9z4"] Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.928389 4944 scope.go:117] "RemoveContainer" containerID="a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b" Nov 24 09:11:43 crc kubenswrapper[4944]: E1124 09:11:43.929541 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b\": container with ID starting with a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b not found: ID does not exist" containerID="a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.929575 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b"} err="failed to get container status \"a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b\": rpc error: code = NotFound desc = could not find container \"a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b\": container with ID starting with a29b904afc0583c0b28bbe729f1898abc49174891492d6c600990ca8a299a30b not found: ID does not exist" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.929595 4944 scope.go:117] "RemoveContainer" containerID="45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37" Nov 24 09:11:43 crc kubenswrapper[4944]: E1124 09:11:43.931028 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37\": container with ID starting with 45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37 not found: ID does not exist" containerID="45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.931066 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37"} err="failed to get container status \"45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37\": rpc error: code = NotFound desc = could not find container \"45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37\": container with ID starting with 45dd3cff9cade4ae036d8fe5a520a2b52bd848101654ceb33a2e1611f76dda37 not found: ID does not exist" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.985437 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.985520 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config-secret\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.985562 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw9rf\" (UniqueName: \"kubernetes.io/projected/9a8dab89-ec08-4497-a890-bcb70c8ea430-kube-api-access-bw9rf\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.985668 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.986662 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.990785 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config-secret\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:43 crc kubenswrapper[4944]: I1124 09:11:43.990915 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.007144 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw9rf\" (UniqueName: \"kubernetes.io/projected/9a8dab89-ec08-4497-a890-bcb70c8ea430-kube-api-access-bw9rf\") pod \"openstackclient\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " pod="openstack/openstackclient" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.081916 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.288176 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d6d9eee-7952-4b40-8958-2048e7cf7a58" path="/var/lib/kubelet/pods/3d6d9eee-7952-4b40-8958-2048e7cf7a58/volumes" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.560712 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 09:11:44 crc kubenswrapper[4944]: W1124 09:11:44.563525 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a8dab89_ec08_4497_a890_bcb70c8ea430.slice/crio-f0888fb3045d88281da186d7e257c308ceb1dccdcf74fe524fc0b7cdbc1258ec WatchSource:0}: Error finding container f0888fb3045d88281da186d7e257c308ceb1dccdcf74fe524fc0b7cdbc1258ec: Status 404 returned error can't find the container with id f0888fb3045d88281da186d7e257c308ceb1dccdcf74fe524fc0b7cdbc1258ec Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.713129 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7bcd854fd7-z2d6k"] Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.721010 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.729526 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.729704 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.735466 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.764138 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7bcd854fd7-z2d6k"] Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.815252 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-config-data\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.815403 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-etc-swift\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.815495 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qtfg\" (UniqueName: \"kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-kube-api-access-4qtfg\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.815550 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-public-tls-certs\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.815663 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-combined-ca-bundle\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.815800 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-log-httpd\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.815867 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-internal-tls-certs\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.815913 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-run-httpd\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.882848 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9a8dab89-ec08-4497-a890-bcb70c8ea430","Type":"ContainerStarted","Data":"f0888fb3045d88281da186d7e257c308ceb1dccdcf74fe524fc0b7cdbc1258ec"} Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.917635 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-config-data\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.917694 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-etc-swift\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.917714 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qtfg\" (UniqueName: \"kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-kube-api-access-4qtfg\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.917741 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-public-tls-certs\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.917773 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-combined-ca-bundle\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.917818 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-log-httpd\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.917849 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-internal-tls-certs\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.917873 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-run-httpd\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.918417 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-run-httpd\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.919359 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-log-httpd\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.924784 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-etc-swift\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.926232 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-public-tls-certs\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.926523 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-combined-ca-bundle\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.926664 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-config-data\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.931626 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-internal-tls-certs\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:44 crc kubenswrapper[4944]: I1124 09:11:44.938838 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qtfg\" (UniqueName: \"kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-kube-api-access-4qtfg\") pod \"swift-proxy-7bcd854fd7-z2d6k\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.058318 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.374331 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.429906 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-combined-ca-bundle\") pod \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.499070 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.499379 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="ceilometer-central-agent" containerID="cri-o://5b9e9df6d119e7e9ed26178412db428957957754d818ba3fd5e628bfb2e730a4" gracePeriod=30 Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.499512 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="proxy-httpd" containerID="cri-o://b322645c4ac77486a9ce3db6ba3bd6419053bba567bd17251759b44dbb6f547e" gracePeriod=30 Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.499531 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c922b8a-f6f5-4801-839c-e67fcef0a67b" (UID: "0c922b8a-f6f5-4801-839c-e67fcef0a67b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.499559 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="sg-core" containerID="cri-o://695684c6c6e8fb9277e670253bd334e3156a4594ad89f89cffb548da35d8d40b" gracePeriod=30 Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.499603 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="ceilometer-notification-agent" containerID="cri-o://e848fb6c2ab71eb4d88e062d7e414fe1e507449e413fc03eb99ca44255bc3f33" gracePeriod=30 Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.510000 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": EOF" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.531549 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-config\") pod \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.531670 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhpsd\" (UniqueName: \"kubernetes.io/projected/0c922b8a-f6f5-4801-839c-e67fcef0a67b-kube-api-access-qhpsd\") pod \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.531748 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-ovndb-tls-certs\") pod \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.531791 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-httpd-config\") pod \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\" (UID: \"0c922b8a-f6f5-4801-839c-e67fcef0a67b\") " Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.532651 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.539182 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "0c922b8a-f6f5-4801-839c-e67fcef0a67b" (UID: "0c922b8a-f6f5-4801-839c-e67fcef0a67b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.540561 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c922b8a-f6f5-4801-839c-e67fcef0a67b-kube-api-access-qhpsd" (OuterVolumeSpecName: "kube-api-access-qhpsd") pod "0c922b8a-f6f5-4801-839c-e67fcef0a67b" (UID: "0c922b8a-f6f5-4801-839c-e67fcef0a67b"). InnerVolumeSpecName "kube-api-access-qhpsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.596177 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-config" (OuterVolumeSpecName: "config") pod "0c922b8a-f6f5-4801-839c-e67fcef0a67b" (UID: "0c922b8a-f6f5-4801-839c-e67fcef0a67b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.634359 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.634400 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhpsd\" (UniqueName: \"kubernetes.io/projected/0c922b8a-f6f5-4801-839c-e67fcef0a67b-kube-api-access-qhpsd\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.634415 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.642183 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "0c922b8a-f6f5-4801-839c-e67fcef0a67b" (UID: "0c922b8a-f6f5-4801-839c-e67fcef0a67b"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.687921 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7bcd854fd7-z2d6k"] Nov 24 09:11:45 crc kubenswrapper[4944]: W1124 09:11:45.698969 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42d1b549_198c_4a0d_a895_2b1f10dc337d.slice/crio-94bb71dd1ae49e891af53f55df27e1e816740e7be55a0f56d91419eaf28f8148 WatchSource:0}: Error finding container 94bb71dd1ae49e891af53f55df27e1e816740e7be55a0f56d91419eaf28f8148: Status 404 returned error can't find the container with id 94bb71dd1ae49e891af53f55df27e1e816740e7be55a0f56d91419eaf28f8148 Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.736608 4944 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c922b8a-f6f5-4801-839c-e67fcef0a67b-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.901782 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" event={"ID":"42d1b549-198c-4a0d-a895-2b1f10dc337d","Type":"ContainerStarted","Data":"94bb71dd1ae49e891af53f55df27e1e816740e7be55a0f56d91419eaf28f8148"} Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.906106 4944 generic.go:334] "Generic (PLEG): container finished" podID="0c922b8a-f6f5-4801-839c-e67fcef0a67b" containerID="02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec" exitCode=0 Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.906284 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f69f9546-4ksss" event={"ID":"0c922b8a-f6f5-4801-839c-e67fcef0a67b","Type":"ContainerDied","Data":"02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec"} Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.906354 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f69f9546-4ksss" event={"ID":"0c922b8a-f6f5-4801-839c-e67fcef0a67b","Type":"ContainerDied","Data":"d0c3fe174959deb0066b04993bb241792858eb342200ba798d19f2f50caaa54e"} Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.906398 4944 scope.go:117] "RemoveContainer" containerID="82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.906309 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54f69f9546-4ksss" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.927475 4944 generic.go:334] "Generic (PLEG): container finished" podID="bb728e49-f81e-4f19-a087-67eda95ff876" containerID="b322645c4ac77486a9ce3db6ba3bd6419053bba567bd17251759b44dbb6f547e" exitCode=0 Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.927513 4944 generic.go:334] "Generic (PLEG): container finished" podID="bb728e49-f81e-4f19-a087-67eda95ff876" containerID="695684c6c6e8fb9277e670253bd334e3156a4594ad89f89cffb548da35d8d40b" exitCode=2 Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.927547 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb728e49-f81e-4f19-a087-67eda95ff876","Type":"ContainerDied","Data":"b322645c4ac77486a9ce3db6ba3bd6419053bba567bd17251759b44dbb6f547e"} Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.927609 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb728e49-f81e-4f19-a087-67eda95ff876","Type":"ContainerDied","Data":"695684c6c6e8fb9277e670253bd334e3156a4594ad89f89cffb548da35d8d40b"} Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.955220 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-54f69f9546-4ksss"] Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.959213 4944 scope.go:117] "RemoveContainer" containerID="02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec" Nov 24 09:11:45 crc kubenswrapper[4944]: I1124 09:11:45.966676 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-54f69f9546-4ksss"] Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.117652 4944 scope.go:117] "RemoveContainer" containerID="82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a" Nov 24 09:11:46 crc kubenswrapper[4944]: E1124 09:11:46.123563 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a\": container with ID starting with 82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a not found: ID does not exist" containerID="82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a" Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.123604 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a"} err="failed to get container status \"82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a\": rpc error: code = NotFound desc = could not find container \"82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a\": container with ID starting with 82ddf04ecf8f35171024f87550c4f57aede33bc31d7bdd98bf9f5af78dc3924a not found: ID does not exist" Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.123630 4944 scope.go:117] "RemoveContainer" containerID="02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec" Nov 24 09:11:46 crc kubenswrapper[4944]: E1124 09:11:46.124221 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec\": container with ID starting with 02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec not found: ID does not exist" containerID="02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec" Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.124262 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec"} err="failed to get container status \"02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec\": rpc error: code = NotFound desc = could not find container \"02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec\": container with ID starting with 02c16642e970a8fbb1baebb3db43998c532ecc521a6811946dcd785e8bfcbcec not found: ID does not exist" Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.294900 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c922b8a-f6f5-4801-839c-e67fcef0a67b" path="/var/lib/kubelet/pods/0c922b8a-f6f5-4801-839c-e67fcef0a67b/volumes" Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.981502 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" event={"ID":"42d1b549-198c-4a0d-a895-2b1f10dc337d","Type":"ContainerStarted","Data":"2c1a579178840a65cb60ad40f3b83a1c327413e6a2d77e8178e33176b65db789"} Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.982142 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" event={"ID":"42d1b549-198c-4a0d-a895-2b1f10dc337d","Type":"ContainerStarted","Data":"b626ffc5d738486bd96a728516fb7e89c1bf6800fb2f82fd0b1bc9e2f6d19268"} Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.985159 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.985179 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.997909 4944 generic.go:334] "Generic (PLEG): container finished" podID="bb728e49-f81e-4f19-a087-67eda95ff876" containerID="e848fb6c2ab71eb4d88e062d7e414fe1e507449e413fc03eb99ca44255bc3f33" exitCode=0 Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.997958 4944 generic.go:334] "Generic (PLEG): container finished" podID="bb728e49-f81e-4f19-a087-67eda95ff876" containerID="5b9e9df6d119e7e9ed26178412db428957957754d818ba3fd5e628bfb2e730a4" exitCode=0 Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.997992 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb728e49-f81e-4f19-a087-67eda95ff876","Type":"ContainerDied","Data":"e848fb6c2ab71eb4d88e062d7e414fe1e507449e413fc03eb99ca44255bc3f33"} Nov 24 09:11:46 crc kubenswrapper[4944]: I1124 09:11:46.998034 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb728e49-f81e-4f19-a087-67eda95ff876","Type":"ContainerDied","Data":"5b9e9df6d119e7e9ed26178412db428957957754d818ba3fd5e628bfb2e730a4"} Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.009274 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" podStartSLOduration=3.009247299 podStartE2EDuration="3.009247299s" podCreationTimestamp="2025-11-24 09:11:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:47.006837382 +0000 UTC m=+1167.541277854" watchObservedRunningTime="2025-11-24 09:11:47.009247299 +0000 UTC m=+1167.543687761" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.141905 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.175271 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bb728e49-f81e-4f19-a087-67eda95ff876" (UID: "bb728e49-f81e-4f19-a087-67eda95ff876"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.174497 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-log-httpd\") pod \"bb728e49-f81e-4f19-a087-67eda95ff876\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.175481 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-sg-core-conf-yaml\") pod \"bb728e49-f81e-4f19-a087-67eda95ff876\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.176672 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-scripts\") pod \"bb728e49-f81e-4f19-a087-67eda95ff876\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.176830 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-config-data\") pod \"bb728e49-f81e-4f19-a087-67eda95ff876\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.176901 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-combined-ca-bundle\") pod \"bb728e49-f81e-4f19-a087-67eda95ff876\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.177018 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94vvf\" (UniqueName: \"kubernetes.io/projected/bb728e49-f81e-4f19-a087-67eda95ff876-kube-api-access-94vvf\") pod \"bb728e49-f81e-4f19-a087-67eda95ff876\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.177167 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-run-httpd\") pod \"bb728e49-f81e-4f19-a087-67eda95ff876\" (UID: \"bb728e49-f81e-4f19-a087-67eda95ff876\") " Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.177807 4944 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.178277 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bb728e49-f81e-4f19-a087-67eda95ff876" (UID: "bb728e49-f81e-4f19-a087-67eda95ff876"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.196936 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-scripts" (OuterVolumeSpecName: "scripts") pod "bb728e49-f81e-4f19-a087-67eda95ff876" (UID: "bb728e49-f81e-4f19-a087-67eda95ff876"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.206941 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb728e49-f81e-4f19-a087-67eda95ff876-kube-api-access-94vvf" (OuterVolumeSpecName: "kube-api-access-94vvf") pod "bb728e49-f81e-4f19-a087-67eda95ff876" (UID: "bb728e49-f81e-4f19-a087-67eda95ff876"). InnerVolumeSpecName "kube-api-access-94vvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.255589 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bb728e49-f81e-4f19-a087-67eda95ff876" (UID: "bb728e49-f81e-4f19-a087-67eda95ff876"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.279171 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94vvf\" (UniqueName: \"kubernetes.io/projected/bb728e49-f81e-4f19-a087-67eda95ff876-kube-api-access-94vvf\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.279374 4944 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bb728e49-f81e-4f19-a087-67eda95ff876-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.279448 4944 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.279502 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.341203 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb728e49-f81e-4f19-a087-67eda95ff876" (UID: "bb728e49-f81e-4f19-a087-67eda95ff876"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.357220 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-config-data" (OuterVolumeSpecName: "config-data") pod "bb728e49-f81e-4f19-a087-67eda95ff876" (UID: "bb728e49-f81e-4f19-a087-67eda95ff876"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.384106 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.384145 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb728e49-f81e-4f19-a087-67eda95ff876-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:47 crc kubenswrapper[4944]: I1124 09:11:47.528590 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.013127 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bb728e49-f81e-4f19-a087-67eda95ff876","Type":"ContainerDied","Data":"58f421e1b79935b6222f2f2f2be71c3f97552ae618693c3307af05a7e39d0a6d"} Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.013199 4944 scope.go:117] "RemoveContainer" containerID="b322645c4ac77486a9ce3db6ba3bd6419053bba567bd17251759b44dbb6f547e" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.014861 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.056971 4944 scope.go:117] "RemoveContainer" containerID="695684c6c6e8fb9277e670253bd334e3156a4594ad89f89cffb548da35d8d40b" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.067995 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.087139 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.102678 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:48 crc kubenswrapper[4944]: E1124 09:11:48.103232 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="proxy-httpd" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103248 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="proxy-httpd" Nov 24 09:11:48 crc kubenswrapper[4944]: E1124 09:11:48.103277 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="ceilometer-notification-agent" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103285 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="ceilometer-notification-agent" Nov 24 09:11:48 crc kubenswrapper[4944]: E1124 09:11:48.103360 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="sg-core" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103368 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="sg-core" Nov 24 09:11:48 crc kubenswrapper[4944]: E1124 09:11:48.103402 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="ceilometer-central-agent" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103409 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="ceilometer-central-agent" Nov 24 09:11:48 crc kubenswrapper[4944]: E1124 09:11:48.103457 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c922b8a-f6f5-4801-839c-e67fcef0a67b" containerName="neutron-httpd" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103465 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c922b8a-f6f5-4801-839c-e67fcef0a67b" containerName="neutron-httpd" Nov 24 09:11:48 crc kubenswrapper[4944]: E1124 09:11:48.103473 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c922b8a-f6f5-4801-839c-e67fcef0a67b" containerName="neutron-api" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103479 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c922b8a-f6f5-4801-839c-e67fcef0a67b" containerName="neutron-api" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103721 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="proxy-httpd" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103757 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="sg-core" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103775 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="ceilometer-notification-agent" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103787 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c922b8a-f6f5-4801-839c-e67fcef0a67b" containerName="neutron-httpd" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103795 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" containerName="ceilometer-central-agent" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.103804 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c922b8a-f6f5-4801-839c-e67fcef0a67b" containerName="neutron-api" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.105619 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.120285 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.120481 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.135071 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.151124 4944 scope.go:117] "RemoveContainer" containerID="e848fb6c2ab71eb4d88e062d7e414fe1e507449e413fc03eb99ca44255bc3f33" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.171493 4944 scope.go:117] "RemoveContainer" containerID="5b9e9df6d119e7e9ed26178412db428957957754d818ba3fd5e628bfb2e730a4" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.198641 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.198694 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-run-httpd\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.198780 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-log-httpd\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.198849 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-config-data\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.199119 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-scripts\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.199314 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.199356 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6swtf\" (UniqueName: \"kubernetes.io/projected/b5703b32-1337-4c0a-b7f2-55715817caf7-kube-api-access-6swtf\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.290094 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb728e49-f81e-4f19-a087-67eda95ff876" path="/var/lib/kubelet/pods/bb728e49-f81e-4f19-a087-67eda95ff876/volumes" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.300610 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.300669 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6swtf\" (UniqueName: \"kubernetes.io/projected/b5703b32-1337-4c0a-b7f2-55715817caf7-kube-api-access-6swtf\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.300693 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.300718 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-run-httpd\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.300758 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-log-httpd\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.300820 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-config-data\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.300956 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-scripts\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.302778 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-run-httpd\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.303501 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-log-httpd\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.307067 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-scripts\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.307381 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-config-data\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.308210 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.316651 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.318815 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6swtf\" (UniqueName: \"kubernetes.io/projected/b5703b32-1337-4c0a-b7f2-55715817caf7-kube-api-access-6swtf\") pod \"ceilometer-0\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.467846 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:11:48 crc kubenswrapper[4944]: I1124 09:11:48.955937 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:48 crc kubenswrapper[4944]: W1124 09:11:48.971295 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5703b32_1337_4c0a_b7f2_55715817caf7.slice/crio-4e06bc5bc162e2288cc9de4cc4e401b47dcb63a8524618b557db608f4ca23d17 WatchSource:0}: Error finding container 4e06bc5bc162e2288cc9de4cc4e401b47dcb63a8524618b557db608f4ca23d17: Status 404 returned error can't find the container with id 4e06bc5bc162e2288cc9de4cc4e401b47dcb63a8524618b557db608f4ca23d17 Nov 24 09:11:49 crc kubenswrapper[4944]: I1124 09:11:49.023358 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5703b32-1337-4c0a-b7f2-55715817caf7","Type":"ContainerStarted","Data":"4e06bc5bc162e2288cc9de4cc4e401b47dcb63a8524618b557db608f4ca23d17"} Nov 24 09:11:49 crc kubenswrapper[4944]: I1124 09:11:49.902214 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:11:50 crc kubenswrapper[4944]: I1124 09:11:50.037836 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5703b32-1337-4c0a-b7f2-55715817caf7","Type":"ContainerStarted","Data":"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f"} Nov 24 09:11:50 crc kubenswrapper[4944]: I1124 09:11:50.067923 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.066910 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.106837 4944 generic.go:334] "Generic (PLEG): container finished" podID="461d6a32-f9db-4344-8f96-737991c925ef" containerID="18258110a4c5ed8692ca22a715d11dcc1f4e93846569f14ad958bb3c2936dec4" exitCode=137 Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.106876 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"461d6a32-f9db-4344-8f96-737991c925ef","Type":"ContainerDied","Data":"18258110a4c5ed8692ca22a715d11dcc1f4e93846569f14ad958bb3c2936dec4"} Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.264481 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.352916 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b6bn\" (UniqueName: \"kubernetes.io/projected/461d6a32-f9db-4344-8f96-737991c925ef-kube-api-access-8b6bn\") pod \"461d6a32-f9db-4344-8f96-737991c925ef\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.352990 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data\") pod \"461d6a32-f9db-4344-8f96-737991c925ef\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.353033 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/461d6a32-f9db-4344-8f96-737991c925ef-etc-machine-id\") pod \"461d6a32-f9db-4344-8f96-737991c925ef\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.353118 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461d6a32-f9db-4344-8f96-737991c925ef-logs\") pod \"461d6a32-f9db-4344-8f96-737991c925ef\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.353167 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-scripts\") pod \"461d6a32-f9db-4344-8f96-737991c925ef\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.353210 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data-custom\") pod \"461d6a32-f9db-4344-8f96-737991c925ef\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.353251 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-combined-ca-bundle\") pod \"461d6a32-f9db-4344-8f96-737991c925ef\" (UID: \"461d6a32-f9db-4344-8f96-737991c925ef\") " Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.355612 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/461d6a32-f9db-4344-8f96-737991c925ef-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "461d6a32-f9db-4344-8f96-737991c925ef" (UID: "461d6a32-f9db-4344-8f96-737991c925ef"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.358332 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/461d6a32-f9db-4344-8f96-737991c925ef-logs" (OuterVolumeSpecName: "logs") pod "461d6a32-f9db-4344-8f96-737991c925ef" (UID: "461d6a32-f9db-4344-8f96-737991c925ef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.362274 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "461d6a32-f9db-4344-8f96-737991c925ef" (UID: "461d6a32-f9db-4344-8f96-737991c925ef"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.365336 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-scripts" (OuterVolumeSpecName: "scripts") pod "461d6a32-f9db-4344-8f96-737991c925ef" (UID: "461d6a32-f9db-4344-8f96-737991c925ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.366205 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/461d6a32-f9db-4344-8f96-737991c925ef-kube-api-access-8b6bn" (OuterVolumeSpecName: "kube-api-access-8b6bn") pod "461d6a32-f9db-4344-8f96-737991c925ef" (UID: "461d6a32-f9db-4344-8f96-737991c925ef"). InnerVolumeSpecName "kube-api-access-8b6bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.397895 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "461d6a32-f9db-4344-8f96-737991c925ef" (UID: "461d6a32-f9db-4344-8f96-737991c925ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.416069 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data" (OuterVolumeSpecName: "config-data") pod "461d6a32-f9db-4344-8f96-737991c925ef" (UID: "461d6a32-f9db-4344-8f96-737991c925ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.455638 4944 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/461d6a32-f9db-4344-8f96-737991c925ef-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.455679 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461d6a32-f9db-4344-8f96-737991c925ef-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.455693 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.455706 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.455718 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.455732 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b6bn\" (UniqueName: \"kubernetes.io/projected/461d6a32-f9db-4344-8f96-737991c925ef-kube-api-access-8b6bn\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:55 crc kubenswrapper[4944]: I1124 09:11:55.455744 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461d6a32-f9db-4344-8f96-737991c925ef-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.116194 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9a8dab89-ec08-4497-a890-bcb70c8ea430","Type":"ContainerStarted","Data":"194effb1be878a12474f5aad321c9b034378e690429268d01ba3188c16709667"} Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.118381 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"461d6a32-f9db-4344-8f96-737991c925ef","Type":"ContainerDied","Data":"e0b2f08216da5d45a7c35682fffb42705bc30f121cd53fa57337453f97c7cb60"} Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.118429 4944 scope.go:117] "RemoveContainer" containerID="18258110a4c5ed8692ca22a715d11dcc1f4e93846569f14ad958bb3c2936dec4" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.118531 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.134118 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5703b32-1337-4c0a-b7f2-55715817caf7","Type":"ContainerStarted","Data":"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829"} Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.134170 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5703b32-1337-4c0a-b7f2-55715817caf7","Type":"ContainerStarted","Data":"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900"} Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.150582 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.729673867 podStartE2EDuration="13.150563766s" podCreationTimestamp="2025-11-24 09:11:43 +0000 UTC" firstStartedPulling="2025-11-24 09:11:44.5660359 +0000 UTC m=+1165.100476362" lastFinishedPulling="2025-11-24 09:11:54.986925799 +0000 UTC m=+1175.521366261" observedRunningTime="2025-11-24 09:11:56.149180331 +0000 UTC m=+1176.683620803" watchObservedRunningTime="2025-11-24 09:11:56.150563766 +0000 UTC m=+1176.685004248" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.155356 4944 scope.go:117] "RemoveContainer" containerID="07a7c5d14487184e00be4545608a72603f1b3f56553d912070d2230c3f22b18c" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.176982 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.191313 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.207797 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:11:56 crc kubenswrapper[4944]: E1124 09:11:56.208159 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="461d6a32-f9db-4344-8f96-737991c925ef" containerName="cinder-api-log" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.208175 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="461d6a32-f9db-4344-8f96-737991c925ef" containerName="cinder-api-log" Nov 24 09:11:56 crc kubenswrapper[4944]: E1124 09:11:56.208207 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="461d6a32-f9db-4344-8f96-737991c925ef" containerName="cinder-api" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.208213 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="461d6a32-f9db-4344-8f96-737991c925ef" containerName="cinder-api" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.208373 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="461d6a32-f9db-4344-8f96-737991c925ef" containerName="cinder-api" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.208395 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="461d6a32-f9db-4344-8f96-737991c925ef" containerName="cinder-api-log" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.209268 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.211600 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.211939 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.220120 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.233846 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.269277 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-scripts\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.269332 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk7nh\" (UniqueName: \"kubernetes.io/projected/b9a8d972-b16f-421e-9fe4-f57653048845-kube-api-access-xk7nh\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.269496 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.269525 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a8d972-b16f-421e-9fe4-f57653048845-logs\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.269635 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.269667 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data-custom\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.269753 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9a8d972-b16f-421e-9fe4-f57653048845-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.269788 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.269916 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.306802 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="461d6a32-f9db-4344-8f96-737991c925ef" path="/var/lib/kubelet/pods/461d6a32-f9db-4344-8f96-737991c925ef/volumes" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.372183 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.372264 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.372397 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-scripts\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.372419 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk7nh\" (UniqueName: \"kubernetes.io/projected/b9a8d972-b16f-421e-9fe4-f57653048845-kube-api-access-xk7nh\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.372538 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.372560 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a8d972-b16f-421e-9fe4-f57653048845-logs\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.372640 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.372675 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data-custom\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.372720 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9a8d972-b16f-421e-9fe4-f57653048845-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.383195 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.390246 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a8d972-b16f-421e-9fe4-f57653048845-logs\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.391145 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9a8d972-b16f-421e-9fe4-f57653048845-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.392566 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.393545 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.405910 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.406499 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-scripts\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.411680 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data-custom\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.429186 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk7nh\" (UniqueName: \"kubernetes.io/projected/b9a8d972-b16f-421e-9fe4-f57653048845-kube-api-access-xk7nh\") pod \"cinder-api-0\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " pod="openstack/cinder-api-0" Nov 24 09:11:56 crc kubenswrapper[4944]: I1124 09:11:56.546432 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 09:11:57 crc kubenswrapper[4944]: I1124 09:11:57.059438 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:11:57 crc kubenswrapper[4944]: W1124 09:11:57.070283 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9a8d972_b16f_421e_9fe4_f57653048845.slice/crio-e6a761e29191681885e80ca6fb5d64dc5f3a481e473545b3ccdabff09f4c9d3a WatchSource:0}: Error finding container e6a761e29191681885e80ca6fb5d64dc5f3a481e473545b3ccdabff09f4c9d3a: Status 404 returned error can't find the container with id e6a761e29191681885e80ca6fb5d64dc5f3a481e473545b3ccdabff09f4c9d3a Nov 24 09:11:57 crc kubenswrapper[4944]: I1124 09:11:57.149908 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a8d972-b16f-421e-9fe4-f57653048845","Type":"ContainerStarted","Data":"e6a761e29191681885e80ca6fb5d64dc5f3a481e473545b3ccdabff09f4c9d3a"} Nov 24 09:11:58 crc kubenswrapper[4944]: I1124 09:11:58.168705 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5703b32-1337-4c0a-b7f2-55715817caf7","Type":"ContainerStarted","Data":"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f"} Nov 24 09:11:58 crc kubenswrapper[4944]: I1124 09:11:58.169315 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 09:11:58 crc kubenswrapper[4944]: I1124 09:11:58.169116 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="ceilometer-central-agent" containerID="cri-o://6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f" gracePeriod=30 Nov 24 09:11:58 crc kubenswrapper[4944]: I1124 09:11:58.169072 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="proxy-httpd" containerID="cri-o://6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f" gracePeriod=30 Nov 24 09:11:58 crc kubenswrapper[4944]: I1124 09:11:58.169121 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="ceilometer-notification-agent" containerID="cri-o://7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900" gracePeriod=30 Nov 24 09:11:58 crc kubenswrapper[4944]: I1124 09:11:58.169104 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="sg-core" containerID="cri-o://3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829" gracePeriod=30 Nov 24 09:11:58 crc kubenswrapper[4944]: I1124 09:11:58.171111 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a8d972-b16f-421e-9fe4-f57653048845","Type":"ContainerStarted","Data":"ef2d54ce03a16a4e0d23c0030984dc6c28b947390104ad9f47f988b5741614d5"} Nov 24 09:11:58 crc kubenswrapper[4944]: I1124 09:11:58.198484 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9583268999999999 podStartE2EDuration="10.198453485s" podCreationTimestamp="2025-11-24 09:11:48 +0000 UTC" firstStartedPulling="2025-11-24 09:11:48.974093021 +0000 UTC m=+1169.508533483" lastFinishedPulling="2025-11-24 09:11:57.214219616 +0000 UTC m=+1177.748660068" observedRunningTime="2025-11-24 09:11:58.193413854 +0000 UTC m=+1178.727854326" watchObservedRunningTime="2025-11-24 09:11:58.198453485 +0000 UTC m=+1178.732893947" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.197104 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a8d972-b16f-421e-9fe4-f57653048845","Type":"ContainerStarted","Data":"84fadfabdc280facbbd9a3442d721a8b2b1c0298a2c5bd828ac25612a83c8d81"} Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.197458 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.198745 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.212391 4944 generic.go:334] "Generic (PLEG): container finished" podID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerID="6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f" exitCode=0 Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.212427 4944 generic.go:334] "Generic (PLEG): container finished" podID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerID="3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829" exitCode=2 Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.212437 4944 generic.go:334] "Generic (PLEG): container finished" podID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerID="7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900" exitCode=0 Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.212445 4944 generic.go:334] "Generic (PLEG): container finished" podID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerID="6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f" exitCode=0 Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.212472 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5703b32-1337-4c0a-b7f2-55715817caf7","Type":"ContainerDied","Data":"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f"} Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.212502 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5703b32-1337-4c0a-b7f2-55715817caf7","Type":"ContainerDied","Data":"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829"} Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.212530 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5703b32-1337-4c0a-b7f2-55715817caf7","Type":"ContainerDied","Data":"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900"} Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.212542 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5703b32-1337-4c0a-b7f2-55715817caf7","Type":"ContainerDied","Data":"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f"} Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.212558 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b5703b32-1337-4c0a-b7f2-55715817caf7","Type":"ContainerDied","Data":"4e06bc5bc162e2288cc9de4cc4e401b47dcb63a8524618b557db608f4ca23d17"} Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.212574 4944 scope.go:117] "RemoveContainer" containerID="6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.230736 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.230707834 podStartE2EDuration="3.230707834s" podCreationTimestamp="2025-11-24 09:11:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:11:59.220995115 +0000 UTC m=+1179.755435587" watchObservedRunningTime="2025-11-24 09:11:59.230707834 +0000 UTC m=+1179.765148296" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.239987 4944 scope.go:117] "RemoveContainer" containerID="3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.266079 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-combined-ca-bundle\") pod \"b5703b32-1337-4c0a-b7f2-55715817caf7\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.266126 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-config-data\") pod \"b5703b32-1337-4c0a-b7f2-55715817caf7\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.266186 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-run-httpd\") pod \"b5703b32-1337-4c0a-b7f2-55715817caf7\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.266225 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6swtf\" (UniqueName: \"kubernetes.io/projected/b5703b32-1337-4c0a-b7f2-55715817caf7-kube-api-access-6swtf\") pod \"b5703b32-1337-4c0a-b7f2-55715817caf7\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.266334 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-log-httpd\") pod \"b5703b32-1337-4c0a-b7f2-55715817caf7\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.266426 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-scripts\") pod \"b5703b32-1337-4c0a-b7f2-55715817caf7\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.266459 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-sg-core-conf-yaml\") pod \"b5703b32-1337-4c0a-b7f2-55715817caf7\" (UID: \"b5703b32-1337-4c0a-b7f2-55715817caf7\") " Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.268100 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b5703b32-1337-4c0a-b7f2-55715817caf7" (UID: "b5703b32-1337-4c0a-b7f2-55715817caf7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.269236 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b5703b32-1337-4c0a-b7f2-55715817caf7" (UID: "b5703b32-1337-4c0a-b7f2-55715817caf7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.286014 4944 scope.go:117] "RemoveContainer" containerID="7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.305072 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5703b32-1337-4c0a-b7f2-55715817caf7-kube-api-access-6swtf" (OuterVolumeSpecName: "kube-api-access-6swtf") pod "b5703b32-1337-4c0a-b7f2-55715817caf7" (UID: "b5703b32-1337-4c0a-b7f2-55715817caf7"). InnerVolumeSpecName "kube-api-access-6swtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.312920 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-scripts" (OuterVolumeSpecName: "scripts") pod "b5703b32-1337-4c0a-b7f2-55715817caf7" (UID: "b5703b32-1337-4c0a-b7f2-55715817caf7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.335479 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b5703b32-1337-4c0a-b7f2-55715817caf7" (UID: "b5703b32-1337-4c0a-b7f2-55715817caf7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.369747 4944 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.369810 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6swtf\" (UniqueName: \"kubernetes.io/projected/b5703b32-1337-4c0a-b7f2-55715817caf7-kube-api-access-6swtf\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.369846 4944 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b5703b32-1337-4c0a-b7f2-55715817caf7-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.369858 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.369868 4944 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.383565 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5703b32-1337-4c0a-b7f2-55715817caf7" (UID: "b5703b32-1337-4c0a-b7f2-55715817caf7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.428802 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-config-data" (OuterVolumeSpecName: "config-data") pod "b5703b32-1337-4c0a-b7f2-55715817caf7" (UID: "b5703b32-1337-4c0a-b7f2-55715817caf7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.441557 4944 scope.go:117] "RemoveContainer" containerID="6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.464271 4944 scope.go:117] "RemoveContainer" containerID="6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f" Nov 24 09:11:59 crc kubenswrapper[4944]: E1124 09:11:59.464698 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f\": container with ID starting with 6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f not found: ID does not exist" containerID="6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.464925 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f"} err="failed to get container status \"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f\": rpc error: code = NotFound desc = could not find container \"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f\": container with ID starting with 6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.464955 4944 scope.go:117] "RemoveContainer" containerID="3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829" Nov 24 09:11:59 crc kubenswrapper[4944]: E1124 09:11:59.465293 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829\": container with ID starting with 3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829 not found: ID does not exist" containerID="3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.465317 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829"} err="failed to get container status \"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829\": rpc error: code = NotFound desc = could not find container \"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829\": container with ID starting with 3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829 not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.465330 4944 scope.go:117] "RemoveContainer" containerID="7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900" Nov 24 09:11:59 crc kubenswrapper[4944]: E1124 09:11:59.465574 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900\": container with ID starting with 7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900 not found: ID does not exist" containerID="7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.465596 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900"} err="failed to get container status \"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900\": rpc error: code = NotFound desc = could not find container \"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900\": container with ID starting with 7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900 not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.465609 4944 scope.go:117] "RemoveContainer" containerID="6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f" Nov 24 09:11:59 crc kubenswrapper[4944]: E1124 09:11:59.466023 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f\": container with ID starting with 6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f not found: ID does not exist" containerID="6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.466098 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f"} err="failed to get container status \"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f\": rpc error: code = NotFound desc = could not find container \"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f\": container with ID starting with 6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.466119 4944 scope.go:117] "RemoveContainer" containerID="6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.466337 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f"} err="failed to get container status \"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f\": rpc error: code = NotFound desc = could not find container \"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f\": container with ID starting with 6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.466357 4944 scope.go:117] "RemoveContainer" containerID="3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.466554 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829"} err="failed to get container status \"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829\": rpc error: code = NotFound desc = could not find container \"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829\": container with ID starting with 3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829 not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.466571 4944 scope.go:117] "RemoveContainer" containerID="7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.466829 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900"} err="failed to get container status \"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900\": rpc error: code = NotFound desc = could not find container \"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900\": container with ID starting with 7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900 not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.466850 4944 scope.go:117] "RemoveContainer" containerID="6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.467169 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f"} err="failed to get container status \"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f\": rpc error: code = NotFound desc = could not find container \"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f\": container with ID starting with 6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.467188 4944 scope.go:117] "RemoveContainer" containerID="6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.467419 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f"} err="failed to get container status \"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f\": rpc error: code = NotFound desc = could not find container \"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f\": container with ID starting with 6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.467434 4944 scope.go:117] "RemoveContainer" containerID="3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.467642 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829"} err="failed to get container status \"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829\": rpc error: code = NotFound desc = could not find container \"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829\": container with ID starting with 3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829 not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.467663 4944 scope.go:117] "RemoveContainer" containerID="7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.467903 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900"} err="failed to get container status \"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900\": rpc error: code = NotFound desc = could not find container \"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900\": container with ID starting with 7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900 not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.467922 4944 scope.go:117] "RemoveContainer" containerID="6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.468101 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f"} err="failed to get container status \"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f\": rpc error: code = NotFound desc = could not find container \"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f\": container with ID starting with 6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.468115 4944 scope.go:117] "RemoveContainer" containerID="6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.468319 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f"} err="failed to get container status \"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f\": rpc error: code = NotFound desc = could not find container \"6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f\": container with ID starting with 6b307ff7758d7ebedb0efc4834e437a1961e4af7e239e165de2c7b82b9599e6f not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.468333 4944 scope.go:117] "RemoveContainer" containerID="3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.468520 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829"} err="failed to get container status \"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829\": rpc error: code = NotFound desc = could not find container \"3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829\": container with ID starting with 3546a8d45bc3bcc637ef41f0e587277a176119b75dd316d2e8de60e895612829 not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.468535 4944 scope.go:117] "RemoveContainer" containerID="7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.468715 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900"} err="failed to get container status \"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900\": rpc error: code = NotFound desc = could not find container \"7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900\": container with ID starting with 7558cb9ff1fcabf5fb40d077712cf89dd812f92cf520a4625e7f8618427ac900 not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.468734 4944 scope.go:117] "RemoveContainer" containerID="6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.468893 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f"} err="failed to get container status \"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f\": rpc error: code = NotFound desc = could not find container \"6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f\": container with ID starting with 6fbd2524e57ede1a25f3ccfe26316bc834c835051c8d2f1cf52436ef8bc8404f not found: ID does not exist" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.471277 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:11:59 crc kubenswrapper[4944]: I1124 09:11:59.471305 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5703b32-1337-4c0a-b7f2-55715817caf7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.222523 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.267212 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.336150 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.336197 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:00 crc kubenswrapper[4944]: E1124 09:12:00.337246 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="sg-core" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.337273 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="sg-core" Nov 24 09:12:00 crc kubenswrapper[4944]: E1124 09:12:00.337311 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="ceilometer-notification-agent" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.337322 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="ceilometer-notification-agent" Nov 24 09:12:00 crc kubenswrapper[4944]: E1124 09:12:00.337345 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="proxy-httpd" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.337353 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="proxy-httpd" Nov 24 09:12:00 crc kubenswrapper[4944]: E1124 09:12:00.337367 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="ceilometer-central-agent" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.337375 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="ceilometer-central-agent" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.337587 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="proxy-httpd" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.337610 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="ceilometer-notification-agent" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.337623 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="sg-core" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.337638 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" containerName="ceilometer-central-agent" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.340250 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.343633 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.343834 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.348235 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.394817 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.394912 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlqwc\" (UniqueName: \"kubernetes.io/projected/5cbd5e05-4fba-4490-be02-8cc6441181cd-kube-api-access-xlqwc\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.395176 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-log-httpd\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.395204 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-run-httpd\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.395229 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-scripts\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.395473 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.395561 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-config-data\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.497463 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.497542 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlqwc\" (UniqueName: \"kubernetes.io/projected/5cbd5e05-4fba-4490-be02-8cc6441181cd-kube-api-access-xlqwc\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.497602 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-log-httpd\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.497642 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-run-httpd\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.497692 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-scripts\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.497719 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.497770 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-config-data\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.498230 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-log-httpd\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.498538 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-run-httpd\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.502691 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-scripts\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.503364 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.503857 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.505489 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-config-data\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.517800 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlqwc\" (UniqueName: \"kubernetes.io/projected/5cbd5e05-4fba-4490-be02-8cc6441181cd-kube-api-access-xlqwc\") pod \"ceilometer-0\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " pod="openstack/ceilometer-0" Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.621668 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:00 crc kubenswrapper[4944]: I1124 09:12:00.622630 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:12:01 crc kubenswrapper[4944]: I1124 09:12:01.067962 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:01 crc kubenswrapper[4944]: W1124 09:12:01.084088 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cbd5e05_4fba_4490_be02_8cc6441181cd.slice/crio-52450ce3d2ed476be62417930f7158ac726f2b669dae59693f4b7944ce6fb2fd WatchSource:0}: Error finding container 52450ce3d2ed476be62417930f7158ac726f2b669dae59693f4b7944ce6fb2fd: Status 404 returned error can't find the container with id 52450ce3d2ed476be62417930f7158ac726f2b669dae59693f4b7944ce6fb2fd Nov 24 09:12:01 crc kubenswrapper[4944]: I1124 09:12:01.232150 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cbd5e05-4fba-4490-be02-8cc6441181cd","Type":"ContainerStarted","Data":"52450ce3d2ed476be62417930f7158ac726f2b669dae59693f4b7944ce6fb2fd"} Nov 24 09:12:02 crc kubenswrapper[4944]: I1124 09:12:02.241545 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cbd5e05-4fba-4490-be02-8cc6441181cd","Type":"ContainerStarted","Data":"137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33"} Nov 24 09:12:02 crc kubenswrapper[4944]: I1124 09:12:02.287283 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5703b32-1337-4c0a-b7f2-55715817caf7" path="/var/lib/kubelet/pods/b5703b32-1337-4c0a-b7f2-55715817caf7/volumes" Nov 24 09:12:03 crc kubenswrapper[4944]: I1124 09:12:03.251826 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cbd5e05-4fba-4490-be02-8cc6441181cd","Type":"ContainerStarted","Data":"2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549"} Nov 24 09:12:03 crc kubenswrapper[4944]: I1124 09:12:03.724858 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:12:03 crc kubenswrapper[4944]: I1124 09:12:03.725444 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1221a90c-370a-4103-bd3c-b3dfe670b446" containerName="glance-log" containerID="cri-o://0d336c6d00011b405b89e69c75750bd4892dffa8ff7440cdf41f4517d10dd129" gracePeriod=30 Nov 24 09:12:03 crc kubenswrapper[4944]: I1124 09:12:03.725581 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1221a90c-370a-4103-bd3c-b3dfe670b446" containerName="glance-httpd" containerID="cri-o://32fb6228e84968d5e793ee6c46d782e11a2ef4eebf6c818f2226f27364162b50" gracePeriod=30 Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.269419 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cbd5e05-4fba-4490-be02-8cc6441181cd","Type":"ContainerStarted","Data":"e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a"} Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.274989 4944 generic.go:334] "Generic (PLEG): container finished" podID="1221a90c-370a-4103-bd3c-b3dfe670b446" containerID="0d336c6d00011b405b89e69c75750bd4892dffa8ff7440cdf41f4517d10dd129" exitCode=143 Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.275066 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1221a90c-370a-4103-bd3c-b3dfe670b446","Type":"ContainerDied","Data":"0d336c6d00011b405b89e69c75750bd4892dffa8ff7440cdf41f4517d10dd129"} Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.675809 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-5bmbf"] Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.677481 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5bmbf" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.686863 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5bmbf"] Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.771246 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-cw4g5"] Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.773305 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk27v\" (UniqueName: \"kubernetes.io/projected/d1ea10a0-65c3-465c-bf40-e1804f588745-kube-api-access-qk27v\") pod \"nova-api-db-create-5bmbf\" (UID: \"d1ea10a0-65c3-465c-bf40-e1804f588745\") " pod="openstack/nova-api-db-create-5bmbf" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.773536 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1ea10a0-65c3-465c-bf40-e1804f588745-operator-scripts\") pod \"nova-api-db-create-5bmbf\" (UID: \"d1ea10a0-65c3-465c-bf40-e1804f588745\") " pod="openstack/nova-api-db-create-5bmbf" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.773840 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cw4g5" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.784586 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cw4g5"] Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.794625 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-e0a2-account-create-xss7d"] Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.796194 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e0a2-account-create-xss7d" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.799206 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.811643 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e0a2-account-create-xss7d"] Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.874985 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1ea10a0-65c3-465c-bf40-e1804f588745-operator-scripts\") pod \"nova-api-db-create-5bmbf\" (UID: \"d1ea10a0-65c3-465c-bf40-e1804f588745\") " pod="openstack/nova-api-db-create-5bmbf" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.875116 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7mkg\" (UniqueName: \"kubernetes.io/projected/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-kube-api-access-d7mkg\") pod \"nova-api-e0a2-account-create-xss7d\" (UID: \"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0\") " pod="openstack/nova-api-e0a2-account-create-xss7d" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.875146 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-operator-scripts\") pod \"nova-api-e0a2-account-create-xss7d\" (UID: \"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0\") " pod="openstack/nova-api-e0a2-account-create-xss7d" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.875203 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk27v\" (UniqueName: \"kubernetes.io/projected/d1ea10a0-65c3-465c-bf40-e1804f588745-kube-api-access-qk27v\") pod \"nova-api-db-create-5bmbf\" (UID: \"d1ea10a0-65c3-465c-bf40-e1804f588745\") " pod="openstack/nova-api-db-create-5bmbf" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.875240 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfb6w\" (UniqueName: \"kubernetes.io/projected/b92b104c-e53e-4246-b5d0-099005284506-kube-api-access-tfb6w\") pod \"nova-cell0-db-create-cw4g5\" (UID: \"b92b104c-e53e-4246-b5d0-099005284506\") " pod="openstack/nova-cell0-db-create-cw4g5" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.875277 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b92b104c-e53e-4246-b5d0-099005284506-operator-scripts\") pod \"nova-cell0-db-create-cw4g5\" (UID: \"b92b104c-e53e-4246-b5d0-099005284506\") " pod="openstack/nova-cell0-db-create-cw4g5" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.876113 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1ea10a0-65c3-465c-bf40-e1804f588745-operator-scripts\") pod \"nova-api-db-create-5bmbf\" (UID: \"d1ea10a0-65c3-465c-bf40-e1804f588745\") " pod="openstack/nova-api-db-create-5bmbf" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.883514 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-d8m52"] Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.885004 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d8m52" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.905612 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-d8m52"] Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.917389 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk27v\" (UniqueName: \"kubernetes.io/projected/d1ea10a0-65c3-465c-bf40-e1804f588745-kube-api-access-qk27v\") pod \"nova-api-db-create-5bmbf\" (UID: \"d1ea10a0-65c3-465c-bf40-e1804f588745\") " pod="openstack/nova-api-db-create-5bmbf" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.976878 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crx7w\" (UniqueName: \"kubernetes.io/projected/60b786e3-e5e4-4503-9887-9a3faef1d5c0-kube-api-access-crx7w\") pod \"nova-cell1-db-create-d8m52\" (UID: \"60b786e3-e5e4-4503-9887-9a3faef1d5c0\") " pod="openstack/nova-cell1-db-create-d8m52" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.976961 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7mkg\" (UniqueName: \"kubernetes.io/projected/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-kube-api-access-d7mkg\") pod \"nova-api-e0a2-account-create-xss7d\" (UID: \"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0\") " pod="openstack/nova-api-e0a2-account-create-xss7d" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.977099 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-operator-scripts\") pod \"nova-api-e0a2-account-create-xss7d\" (UID: \"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0\") " pod="openstack/nova-api-e0a2-account-create-xss7d" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.977429 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfb6w\" (UniqueName: \"kubernetes.io/projected/b92b104c-e53e-4246-b5d0-099005284506-kube-api-access-tfb6w\") pod \"nova-cell0-db-create-cw4g5\" (UID: \"b92b104c-e53e-4246-b5d0-099005284506\") " pod="openstack/nova-cell0-db-create-cw4g5" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.977473 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b786e3-e5e4-4503-9887-9a3faef1d5c0-operator-scripts\") pod \"nova-cell1-db-create-d8m52\" (UID: \"60b786e3-e5e4-4503-9887-9a3faef1d5c0\") " pod="openstack/nova-cell1-db-create-d8m52" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.977529 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b92b104c-e53e-4246-b5d0-099005284506-operator-scripts\") pod \"nova-cell0-db-create-cw4g5\" (UID: \"b92b104c-e53e-4246-b5d0-099005284506\") " pod="openstack/nova-cell0-db-create-cw4g5" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.977968 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-operator-scripts\") pod \"nova-api-e0a2-account-create-xss7d\" (UID: \"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0\") " pod="openstack/nova-api-e0a2-account-create-xss7d" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.978297 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b92b104c-e53e-4246-b5d0-099005284506-operator-scripts\") pod \"nova-cell0-db-create-cw4g5\" (UID: \"b92b104c-e53e-4246-b5d0-099005284506\") " pod="openstack/nova-cell0-db-create-cw4g5" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.981502 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-eec8-account-create-snmtp"] Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.982621 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-eec8-account-create-snmtp" Nov 24 09:12:04 crc kubenswrapper[4944]: I1124 09:12:04.984909 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.001296 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfb6w\" (UniqueName: \"kubernetes.io/projected/b92b104c-e53e-4246-b5d0-099005284506-kube-api-access-tfb6w\") pod \"nova-cell0-db-create-cw4g5\" (UID: \"b92b104c-e53e-4246-b5d0-099005284506\") " pod="openstack/nova-cell0-db-create-cw4g5" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.003750 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7mkg\" (UniqueName: \"kubernetes.io/projected/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-kube-api-access-d7mkg\") pod \"nova-api-e0a2-account-create-xss7d\" (UID: \"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0\") " pod="openstack/nova-api-e0a2-account-create-xss7d" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.027671 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-eec8-account-create-snmtp"] Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.029281 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5bmbf" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.079374 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b786e3-e5e4-4503-9887-9a3faef1d5c0-operator-scripts\") pod \"nova-cell1-db-create-d8m52\" (UID: \"60b786e3-e5e4-4503-9887-9a3faef1d5c0\") " pod="openstack/nova-cell1-db-create-d8m52" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.079464 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b935dbba-ca03-419e-95de-f1061fbc0999-operator-scripts\") pod \"nova-cell0-eec8-account-create-snmtp\" (UID: \"b935dbba-ca03-419e-95de-f1061fbc0999\") " pod="openstack/nova-cell0-eec8-account-create-snmtp" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.079516 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crx7w\" (UniqueName: \"kubernetes.io/projected/60b786e3-e5e4-4503-9887-9a3faef1d5c0-kube-api-access-crx7w\") pod \"nova-cell1-db-create-d8m52\" (UID: \"60b786e3-e5e4-4503-9887-9a3faef1d5c0\") " pod="openstack/nova-cell1-db-create-d8m52" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.079569 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbfmx\" (UniqueName: \"kubernetes.io/projected/b935dbba-ca03-419e-95de-f1061fbc0999-kube-api-access-hbfmx\") pod \"nova-cell0-eec8-account-create-snmtp\" (UID: \"b935dbba-ca03-419e-95de-f1061fbc0999\") " pod="openstack/nova-cell0-eec8-account-create-snmtp" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.080474 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b786e3-e5e4-4503-9887-9a3faef1d5c0-operator-scripts\") pod \"nova-cell1-db-create-d8m52\" (UID: \"60b786e3-e5e4-4503-9887-9a3faef1d5c0\") " pod="openstack/nova-cell1-db-create-d8m52" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.092735 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cw4g5" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.114895 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crx7w\" (UniqueName: \"kubernetes.io/projected/60b786e3-e5e4-4503-9887-9a3faef1d5c0-kube-api-access-crx7w\") pod \"nova-cell1-db-create-d8m52\" (UID: \"60b786e3-e5e4-4503-9887-9a3faef1d5c0\") " pod="openstack/nova-cell1-db-create-d8m52" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.121516 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e0a2-account-create-xss7d" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.151254 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.159339 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="89e12262-3506-49db-b83c-e9a53974ad5d" containerName="glance-log" containerID="cri-o://8bda4061dc19bf95a403f1262f4d42e018ccabe08f270939e7945ba45cb194b9" gracePeriod=30 Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.159512 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="89e12262-3506-49db-b83c-e9a53974ad5d" containerName="glance-httpd" containerID="cri-o://d3c3b5c16d1ed60f1fdf53659af98f1126eb3740de0da4f9ea843419e7af7ba9" gracePeriod=30 Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.184102 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbfmx\" (UniqueName: \"kubernetes.io/projected/b935dbba-ca03-419e-95de-f1061fbc0999-kube-api-access-hbfmx\") pod \"nova-cell0-eec8-account-create-snmtp\" (UID: \"b935dbba-ca03-419e-95de-f1061fbc0999\") " pod="openstack/nova-cell0-eec8-account-create-snmtp" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.184232 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b935dbba-ca03-419e-95de-f1061fbc0999-operator-scripts\") pod \"nova-cell0-eec8-account-create-snmtp\" (UID: \"b935dbba-ca03-419e-95de-f1061fbc0999\") " pod="openstack/nova-cell0-eec8-account-create-snmtp" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.184994 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b935dbba-ca03-419e-95de-f1061fbc0999-operator-scripts\") pod \"nova-cell0-eec8-account-create-snmtp\" (UID: \"b935dbba-ca03-419e-95de-f1061fbc0999\") " pod="openstack/nova-cell0-eec8-account-create-snmtp" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.210639 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbfmx\" (UniqueName: \"kubernetes.io/projected/b935dbba-ca03-419e-95de-f1061fbc0999-kube-api-access-hbfmx\") pod \"nova-cell0-eec8-account-create-snmtp\" (UID: \"b935dbba-ca03-419e-95de-f1061fbc0999\") " pod="openstack/nova-cell0-eec8-account-create-snmtp" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.227765 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-2d40-account-create-l7jv5"] Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.229097 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2d40-account-create-l7jv5" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.229505 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d8m52" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.235003 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.245657 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2d40-account-create-l7jv5"] Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.287025 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgv78\" (UniqueName: \"kubernetes.io/projected/5ae35e54-8396-4a44-90d7-ae4afd3322af-kube-api-access-zgv78\") pod \"nova-cell1-2d40-account-create-l7jv5\" (UID: \"5ae35e54-8396-4a44-90d7-ae4afd3322af\") " pod="openstack/nova-cell1-2d40-account-create-l7jv5" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.287158 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ae35e54-8396-4a44-90d7-ae4afd3322af-operator-scripts\") pod \"nova-cell1-2d40-account-create-l7jv5\" (UID: \"5ae35e54-8396-4a44-90d7-ae4afd3322af\") " pod="openstack/nova-cell1-2d40-account-create-l7jv5" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.302562 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-eec8-account-create-snmtp" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.391028 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgv78\" (UniqueName: \"kubernetes.io/projected/5ae35e54-8396-4a44-90d7-ae4afd3322af-kube-api-access-zgv78\") pod \"nova-cell1-2d40-account-create-l7jv5\" (UID: \"5ae35e54-8396-4a44-90d7-ae4afd3322af\") " pod="openstack/nova-cell1-2d40-account-create-l7jv5" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.391219 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ae35e54-8396-4a44-90d7-ae4afd3322af-operator-scripts\") pod \"nova-cell1-2d40-account-create-l7jv5\" (UID: \"5ae35e54-8396-4a44-90d7-ae4afd3322af\") " pod="openstack/nova-cell1-2d40-account-create-l7jv5" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.392538 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ae35e54-8396-4a44-90d7-ae4afd3322af-operator-scripts\") pod \"nova-cell1-2d40-account-create-l7jv5\" (UID: \"5ae35e54-8396-4a44-90d7-ae4afd3322af\") " pod="openstack/nova-cell1-2d40-account-create-l7jv5" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.415896 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cbd5e05-4fba-4490-be02-8cc6441181cd","Type":"ContainerStarted","Data":"b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c"} Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.416106 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="ceilometer-central-agent" containerID="cri-o://137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33" gracePeriod=30 Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.416415 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.416642 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="proxy-httpd" containerID="cri-o://b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c" gracePeriod=30 Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.416661 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="sg-core" containerID="cri-o://e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a" gracePeriod=30 Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.416720 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="ceilometer-notification-agent" containerID="cri-o://2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549" gracePeriod=30 Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.462491 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.981323889 podStartE2EDuration="5.462471909s" podCreationTimestamp="2025-11-24 09:12:00 +0000 UTC" firstStartedPulling="2025-11-24 09:12:01.087579025 +0000 UTC m=+1181.622019487" lastFinishedPulling="2025-11-24 09:12:04.568727035 +0000 UTC m=+1185.103167507" observedRunningTime="2025-11-24 09:12:05.460975672 +0000 UTC m=+1185.995416134" watchObservedRunningTime="2025-11-24 09:12:05.462471909 +0000 UTC m=+1185.996912371" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.467721 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgv78\" (UniqueName: \"kubernetes.io/projected/5ae35e54-8396-4a44-90d7-ae4afd3322af-kube-api-access-zgv78\") pod \"nova-cell1-2d40-account-create-l7jv5\" (UID: \"5ae35e54-8396-4a44-90d7-ae4afd3322af\") " pod="openstack/nova-cell1-2d40-account-create-l7jv5" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.577083 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2d40-account-create-l7jv5" Nov 24 09:12:05 crc kubenswrapper[4944]: I1124 09:12:05.697694 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5bmbf"] Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.076964 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e0a2-account-create-xss7d"] Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.200431 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-d8m52"] Nov 24 09:12:06 crc kubenswrapper[4944]: W1124 09:12:06.213843 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb935dbba_ca03_419e_95de_f1061fbc0999.slice/crio-a2c953af2c37256708ceaff813f7bf5d47356f7668188ddf42483aee0931adfc WatchSource:0}: Error finding container a2c953af2c37256708ceaff813f7bf5d47356f7668188ddf42483aee0931adfc: Status 404 returned error can't find the container with id a2c953af2c37256708ceaff813f7bf5d47356f7668188ddf42483aee0931adfc Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.228025 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cw4g5"] Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.241561 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-eec8-account-create-snmtp"] Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.397342 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2d40-account-create-l7jv5"] Nov 24 09:12:06 crc kubenswrapper[4944]: W1124 09:12:06.414346 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ae35e54_8396_4a44_90d7_ae4afd3322af.slice/crio-abdee8457052c756ad740afc7cf6414bb45826ad5b64991fd5fd1b17614b4820 WatchSource:0}: Error finding container abdee8457052c756ad740afc7cf6414bb45826ad5b64991fd5fd1b17614b4820: Status 404 returned error can't find the container with id abdee8457052c756ad740afc7cf6414bb45826ad5b64991fd5fd1b17614b4820 Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.429347 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2d40-account-create-l7jv5" event={"ID":"5ae35e54-8396-4a44-90d7-ae4afd3322af","Type":"ContainerStarted","Data":"abdee8457052c756ad740afc7cf6414bb45826ad5b64991fd5fd1b17614b4820"} Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.433750 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-d8m52" event={"ID":"60b786e3-e5e4-4503-9887-9a3faef1d5c0","Type":"ContainerStarted","Data":"a48a5b30be83ad437a2d8550494ed3e5b2ff94ac6eeb271658df43bee55c4bc5"} Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.438018 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cw4g5" event={"ID":"b92b104c-e53e-4246-b5d0-099005284506","Type":"ContainerStarted","Data":"dd2f68edf9b5d85498f177bfea600649f2d1a5dd919f99b7ad318e69e1bf167f"} Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.441495 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e0a2-account-create-xss7d" event={"ID":"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0","Type":"ContainerStarted","Data":"767f2348215a12909125bab4024992f9be9ed81b2df539cb5b14d05f21908c76"} Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.448219 4944 generic.go:334] "Generic (PLEG): container finished" podID="d1ea10a0-65c3-465c-bf40-e1804f588745" containerID="8e5045f0585e8dd46160e5e97aae1809dc8d19296f6529d8d76bb17c76758990" exitCode=0 Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.448328 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5bmbf" event={"ID":"d1ea10a0-65c3-465c-bf40-e1804f588745","Type":"ContainerDied","Data":"8e5045f0585e8dd46160e5e97aae1809dc8d19296f6529d8d76bb17c76758990"} Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.448392 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5bmbf" event={"ID":"d1ea10a0-65c3-465c-bf40-e1804f588745","Type":"ContainerStarted","Data":"5141c696e8818384917e9206ad89c887d8145dd5c5c377ea95c79f9613db306b"} Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.450573 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-eec8-account-create-snmtp" event={"ID":"b935dbba-ca03-419e-95de-f1061fbc0999","Type":"ContainerStarted","Data":"a2c953af2c37256708ceaff813f7bf5d47356f7668188ddf42483aee0931adfc"} Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.458287 4944 generic.go:334] "Generic (PLEG): container finished" podID="89e12262-3506-49db-b83c-e9a53974ad5d" containerID="8bda4061dc19bf95a403f1262f4d42e018ccabe08f270939e7945ba45cb194b9" exitCode=143 Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.458371 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"89e12262-3506-49db-b83c-e9a53974ad5d","Type":"ContainerDied","Data":"8bda4061dc19bf95a403f1262f4d42e018ccabe08f270939e7945ba45cb194b9"} Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.473580 4944 generic.go:334] "Generic (PLEG): container finished" podID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerID="b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c" exitCode=0 Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.473622 4944 generic.go:334] "Generic (PLEG): container finished" podID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerID="e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a" exitCode=2 Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.473634 4944 generic.go:334] "Generic (PLEG): container finished" podID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerID="2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549" exitCode=0 Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.473661 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cbd5e05-4fba-4490-be02-8cc6441181cd","Type":"ContainerDied","Data":"b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c"} Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.473690 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cbd5e05-4fba-4490-be02-8cc6441181cd","Type":"ContainerDied","Data":"e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a"} Nov 24 09:12:06 crc kubenswrapper[4944]: I1124 09:12:06.473702 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cbd5e05-4fba-4490-be02-8cc6441181cd","Type":"ContainerDied","Data":"2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549"} Nov 24 09:12:06 crc kubenswrapper[4944]: E1124 09:12:06.818341 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc7dc25c_e0b6_4dc8_92f4_975dd74346a0.slice/crio-conmon-5c74c827e2500bbe5cec629901b4499f9820b0c3015d1b7fc182c6d47192e91d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc7dc25c_e0b6_4dc8_92f4_975dd74346a0.slice/crio-5c74c827e2500bbe5cec629901b4499f9820b0c3015d1b7fc182c6d47192e91d.scope\": RecentStats: unable to find data in memory cache]" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.486085 4944 generic.go:334] "Generic (PLEG): container finished" podID="b935dbba-ca03-419e-95de-f1061fbc0999" containerID="38bd22ea8e5f5357cd4719266b4150ee62593a72993379b83a9ba97d0c889479" exitCode=0 Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.486167 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-eec8-account-create-snmtp" event={"ID":"b935dbba-ca03-419e-95de-f1061fbc0999","Type":"ContainerDied","Data":"38bd22ea8e5f5357cd4719266b4150ee62593a72993379b83a9ba97d0c889479"} Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.493949 4944 generic.go:334] "Generic (PLEG): container finished" podID="5ae35e54-8396-4a44-90d7-ae4afd3322af" containerID="7fb724f45219e9d1f3b1a09039899d4d0f3f225aabff2cb17b1ae874531f5a97" exitCode=0 Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.494014 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2d40-account-create-l7jv5" event={"ID":"5ae35e54-8396-4a44-90d7-ae4afd3322af","Type":"ContainerDied","Data":"7fb724f45219e9d1f3b1a09039899d4d0f3f225aabff2cb17b1ae874531f5a97"} Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.495930 4944 generic.go:334] "Generic (PLEG): container finished" podID="60b786e3-e5e4-4503-9887-9a3faef1d5c0" containerID="8ee523b44e35b1cfe2ffff334cb3b42352d07e0931e975529ab8ff087be0b252" exitCode=0 Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.495996 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-d8m52" event={"ID":"60b786e3-e5e4-4503-9887-9a3faef1d5c0","Type":"ContainerDied","Data":"8ee523b44e35b1cfe2ffff334cb3b42352d07e0931e975529ab8ff087be0b252"} Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.497230 4944 generic.go:334] "Generic (PLEG): container finished" podID="b92b104c-e53e-4246-b5d0-099005284506" containerID="ba5de3bc08fb3adcb30a032e46fa2306b0ea56b6eef98ca20a635459d7ab466d" exitCode=0 Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.497275 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cw4g5" event={"ID":"b92b104c-e53e-4246-b5d0-099005284506","Type":"ContainerDied","Data":"ba5de3bc08fb3adcb30a032e46fa2306b0ea56b6eef98ca20a635459d7ab466d"} Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.498403 4944 generic.go:334] "Generic (PLEG): container finished" podID="dc7dc25c-e0b6-4dc8-92f4-975dd74346a0" containerID="5c74c827e2500bbe5cec629901b4499f9820b0c3015d1b7fc182c6d47192e91d" exitCode=0 Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.498445 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e0a2-account-create-xss7d" event={"ID":"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0","Type":"ContainerDied","Data":"5c74c827e2500bbe5cec629901b4499f9820b0c3015d1b7fc182c6d47192e91d"} Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.503034 4944 generic.go:334] "Generic (PLEG): container finished" podID="1221a90c-370a-4103-bd3c-b3dfe670b446" containerID="32fb6228e84968d5e793ee6c46d782e11a2ef4eebf6c818f2226f27364162b50" exitCode=0 Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.503071 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1221a90c-370a-4103-bd3c-b3dfe670b446","Type":"ContainerDied","Data":"32fb6228e84968d5e793ee6c46d782e11a2ef4eebf6c818f2226f27364162b50"} Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.503099 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1221a90c-370a-4103-bd3c-b3dfe670b446","Type":"ContainerDied","Data":"5fd55c50b1f605f2526663a959c2eef7286669d0dc1d75b76ac68a84af2bc045"} Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.503113 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fd55c50b1f605f2526663a959c2eef7286669d0dc1d75b76ac68a84af2bc045" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.527169 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.660687 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-scripts\") pod \"1221a90c-370a-4103-bd3c-b3dfe670b446\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.660787 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc86b\" (UniqueName: \"kubernetes.io/projected/1221a90c-370a-4103-bd3c-b3dfe670b446-kube-api-access-lc86b\") pod \"1221a90c-370a-4103-bd3c-b3dfe670b446\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.660837 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"1221a90c-370a-4103-bd3c-b3dfe670b446\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.660887 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-config-data\") pod \"1221a90c-370a-4103-bd3c-b3dfe670b446\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.660973 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-logs\") pod \"1221a90c-370a-4103-bd3c-b3dfe670b446\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.661078 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-combined-ca-bundle\") pod \"1221a90c-370a-4103-bd3c-b3dfe670b446\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.661124 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-httpd-run\") pod \"1221a90c-370a-4103-bd3c-b3dfe670b446\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.661221 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-public-tls-certs\") pod \"1221a90c-370a-4103-bd3c-b3dfe670b446\" (UID: \"1221a90c-370a-4103-bd3c-b3dfe670b446\") " Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.667676 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1221a90c-370a-4103-bd3c-b3dfe670b446-kube-api-access-lc86b" (OuterVolumeSpecName: "kube-api-access-lc86b") pod "1221a90c-370a-4103-bd3c-b3dfe670b446" (UID: "1221a90c-370a-4103-bd3c-b3dfe670b446"). InnerVolumeSpecName "kube-api-access-lc86b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.667838 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-logs" (OuterVolumeSpecName: "logs") pod "1221a90c-370a-4103-bd3c-b3dfe670b446" (UID: "1221a90c-370a-4103-bd3c-b3dfe670b446"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.669236 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "1221a90c-370a-4103-bd3c-b3dfe670b446" (UID: "1221a90c-370a-4103-bd3c-b3dfe670b446"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.673362 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1221a90c-370a-4103-bd3c-b3dfe670b446" (UID: "1221a90c-370a-4103-bd3c-b3dfe670b446"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.706235 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-scripts" (OuterVolumeSpecName: "scripts") pod "1221a90c-370a-4103-bd3c-b3dfe670b446" (UID: "1221a90c-370a-4103-bd3c-b3dfe670b446"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.725116 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-config-data" (OuterVolumeSpecName: "config-data") pod "1221a90c-370a-4103-bd3c-b3dfe670b446" (UID: "1221a90c-370a-4103-bd3c-b3dfe670b446"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.762690 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc86b\" (UniqueName: \"kubernetes.io/projected/1221a90c-370a-4103-bd3c-b3dfe670b446-kube-api-access-lc86b\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.762736 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.762745 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.762754 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.762763 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1221a90c-370a-4103-bd3c-b3dfe670b446-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.762771 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.772356 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1221a90c-370a-4103-bd3c-b3dfe670b446" (UID: "1221a90c-370a-4103-bd3c-b3dfe670b446"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.772465 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1221a90c-370a-4103-bd3c-b3dfe670b446" (UID: "1221a90c-370a-4103-bd3c-b3dfe670b446"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.818834 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.849405 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5bmbf" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.870016 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.870069 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.870080 4944 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1221a90c-370a-4103-bd3c-b3dfe670b446-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.971083 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1ea10a0-65c3-465c-bf40-e1804f588745-operator-scripts\") pod \"d1ea10a0-65c3-465c-bf40-e1804f588745\" (UID: \"d1ea10a0-65c3-465c-bf40-e1804f588745\") " Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.971435 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk27v\" (UniqueName: \"kubernetes.io/projected/d1ea10a0-65c3-465c-bf40-e1804f588745-kube-api-access-qk27v\") pod \"d1ea10a0-65c3-465c-bf40-e1804f588745\" (UID: \"d1ea10a0-65c3-465c-bf40-e1804f588745\") " Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.971768 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1ea10a0-65c3-465c-bf40-e1804f588745-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d1ea10a0-65c3-465c-bf40-e1804f588745" (UID: "d1ea10a0-65c3-465c-bf40-e1804f588745"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.972561 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1ea10a0-65c3-465c-bf40-e1804f588745-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:07 crc kubenswrapper[4944]: I1124 09:12:07.976279 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1ea10a0-65c3-465c-bf40-e1804f588745-kube-api-access-qk27v" (OuterVolumeSpecName: "kube-api-access-qk27v") pod "d1ea10a0-65c3-465c-bf40-e1804f588745" (UID: "d1ea10a0-65c3-465c-bf40-e1804f588745"). InnerVolumeSpecName "kube-api-access-qk27v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.075097 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk27v\" (UniqueName: \"kubernetes.io/projected/d1ea10a0-65c3-465c-bf40-e1804f588745-kube-api-access-qk27v\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.513782 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5bmbf" event={"ID":"d1ea10a0-65c3-465c-bf40-e1804f588745","Type":"ContainerDied","Data":"5141c696e8818384917e9206ad89c887d8145dd5c5c377ea95c79f9613db306b"} Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.513874 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5141c696e8818384917e9206ad89c887d8145dd5c5c377ea95c79f9613db306b" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.513997 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5bmbf" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.514122 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.559113 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.581411 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.623158 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:12:08 crc kubenswrapper[4944]: E1124 09:12:08.623626 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1221a90c-370a-4103-bd3c-b3dfe670b446" containerName="glance-log" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.623645 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1221a90c-370a-4103-bd3c-b3dfe670b446" containerName="glance-log" Nov 24 09:12:08 crc kubenswrapper[4944]: E1124 09:12:08.623669 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ea10a0-65c3-465c-bf40-e1804f588745" containerName="mariadb-database-create" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.623679 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ea10a0-65c3-465c-bf40-e1804f588745" containerName="mariadb-database-create" Nov 24 09:12:08 crc kubenswrapper[4944]: E1124 09:12:08.623708 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1221a90c-370a-4103-bd3c-b3dfe670b446" containerName="glance-httpd" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.623715 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1221a90c-370a-4103-bd3c-b3dfe670b446" containerName="glance-httpd" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.623933 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1ea10a0-65c3-465c-bf40-e1804f588745" containerName="mariadb-database-create" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.623963 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1221a90c-370a-4103-bd3c-b3dfe670b446" containerName="glance-log" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.623978 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1221a90c-370a-4103-bd3c-b3dfe670b446" containerName="glance-httpd" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.625132 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.629439 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.629613 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.640664 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.788598 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-scripts\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.788704 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.788777 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.788873 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.788924 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p56z\" (UniqueName: \"kubernetes.io/projected/f94b9960-fa37-49f7-9fa7-0e01f855992b-kube-api-access-9p56z\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.788956 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.789037 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-logs\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.789232 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-config-data\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.891965 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-logs\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.899959 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-config-data\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.900080 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-scripts\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.900110 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.900173 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.900288 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.900318 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p56z\" (UniqueName: \"kubernetes.io/projected/f94b9960-fa37-49f7-9fa7-0e01f855992b-kube-api-access-9p56z\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.900354 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.902008 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.892643 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-logs\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.910132 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.913716 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.915662 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-scripts\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.915887 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.926410 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-config-data\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:08 crc kubenswrapper[4944]: I1124 09:12:08.932962 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p56z\" (UniqueName: \"kubernetes.io/projected/f94b9960-fa37-49f7-9fa7-0e01f855992b-kube-api-access-9p56z\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.004348 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " pod="openstack/glance-default-external-api-0" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.123453 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d8m52" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.186894 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.258368 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.315720 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b786e3-e5e4-4503-9887-9a3faef1d5c0-operator-scripts\") pod \"60b786e3-e5e4-4503-9887-9a3faef1d5c0\" (UID: \"60b786e3-e5e4-4503-9887-9a3faef1d5c0\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.316227 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crx7w\" (UniqueName: \"kubernetes.io/projected/60b786e3-e5e4-4503-9887-9a3faef1d5c0-kube-api-access-crx7w\") pod \"60b786e3-e5e4-4503-9887-9a3faef1d5c0\" (UID: \"60b786e3-e5e4-4503-9887-9a3faef1d5c0\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.317063 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60b786e3-e5e4-4503-9887-9a3faef1d5c0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60b786e3-e5e4-4503-9887-9a3faef1d5c0" (UID: "60b786e3-e5e4-4503-9887-9a3faef1d5c0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.336326 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60b786e3-e5e4-4503-9887-9a3faef1d5c0-kube-api-access-crx7w" (OuterVolumeSpecName: "kube-api-access-crx7w") pod "60b786e3-e5e4-4503-9887-9a3faef1d5c0" (UID: "60b786e3-e5e4-4503-9887-9a3faef1d5c0"). InnerVolumeSpecName "kube-api-access-crx7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.348849 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e0a2-account-create-xss7d" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.357877 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-eec8-account-create-snmtp" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.370138 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cw4g5" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.378406 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2d40-account-create-l7jv5" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.418367 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b786e3-e5e4-4503-9887-9a3faef1d5c0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.418399 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crx7w\" (UniqueName: \"kubernetes.io/projected/60b786e3-e5e4-4503-9887-9a3faef1d5c0-kube-api-access-crx7w\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.519289 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgv78\" (UniqueName: \"kubernetes.io/projected/5ae35e54-8396-4a44-90d7-ae4afd3322af-kube-api-access-zgv78\") pod \"5ae35e54-8396-4a44-90d7-ae4afd3322af\" (UID: \"5ae35e54-8396-4a44-90d7-ae4afd3322af\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.519333 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b92b104c-e53e-4246-b5d0-099005284506-operator-scripts\") pod \"b92b104c-e53e-4246-b5d0-099005284506\" (UID: \"b92b104c-e53e-4246-b5d0-099005284506\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.519415 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ae35e54-8396-4a44-90d7-ae4afd3322af-operator-scripts\") pod \"5ae35e54-8396-4a44-90d7-ae4afd3322af\" (UID: \"5ae35e54-8396-4a44-90d7-ae4afd3322af\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.519499 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfb6w\" (UniqueName: \"kubernetes.io/projected/b92b104c-e53e-4246-b5d0-099005284506-kube-api-access-tfb6w\") pod \"b92b104c-e53e-4246-b5d0-099005284506\" (UID: \"b92b104c-e53e-4246-b5d0-099005284506\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.519556 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b935dbba-ca03-419e-95de-f1061fbc0999-operator-scripts\") pod \"b935dbba-ca03-419e-95de-f1061fbc0999\" (UID: \"b935dbba-ca03-419e-95de-f1061fbc0999\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.519623 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbfmx\" (UniqueName: \"kubernetes.io/projected/b935dbba-ca03-419e-95de-f1061fbc0999-kube-api-access-hbfmx\") pod \"b935dbba-ca03-419e-95de-f1061fbc0999\" (UID: \"b935dbba-ca03-419e-95de-f1061fbc0999\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.519669 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7mkg\" (UniqueName: \"kubernetes.io/projected/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-kube-api-access-d7mkg\") pod \"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0\" (UID: \"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.519687 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-operator-scripts\") pod \"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0\" (UID: \"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.520874 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b92b104c-e53e-4246-b5d0-099005284506-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b92b104c-e53e-4246-b5d0-099005284506" (UID: "b92b104c-e53e-4246-b5d0-099005284506"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.520981 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b935dbba-ca03-419e-95de-f1061fbc0999-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b935dbba-ca03-419e-95de-f1061fbc0999" (UID: "b935dbba-ca03-419e-95de-f1061fbc0999"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.521337 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b935dbba-ca03-419e-95de-f1061fbc0999-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.521358 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b92b104c-e53e-4246-b5d0-099005284506-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.521394 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ae35e54-8396-4a44-90d7-ae4afd3322af-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ae35e54-8396-4a44-90d7-ae4afd3322af" (UID: "5ae35e54-8396-4a44-90d7-ae4afd3322af"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.521812 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc7dc25c-e0b6-4dc8-92f4-975dd74346a0" (UID: "dc7dc25c-e0b6-4dc8-92f4-975dd74346a0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.524623 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b92b104c-e53e-4246-b5d0-099005284506-kube-api-access-tfb6w" (OuterVolumeSpecName: "kube-api-access-tfb6w") pod "b92b104c-e53e-4246-b5d0-099005284506" (UID: "b92b104c-e53e-4246-b5d0-099005284506"). InnerVolumeSpecName "kube-api-access-tfb6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.526469 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-kube-api-access-d7mkg" (OuterVolumeSpecName: "kube-api-access-d7mkg") pod "dc7dc25c-e0b6-4dc8-92f4-975dd74346a0" (UID: "dc7dc25c-e0b6-4dc8-92f4-975dd74346a0"). InnerVolumeSpecName "kube-api-access-d7mkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.526563 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b935dbba-ca03-419e-95de-f1061fbc0999-kube-api-access-hbfmx" (OuterVolumeSpecName: "kube-api-access-hbfmx") pod "b935dbba-ca03-419e-95de-f1061fbc0999" (UID: "b935dbba-ca03-419e-95de-f1061fbc0999"). InnerVolumeSpecName "kube-api-access-hbfmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.527125 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae35e54-8396-4a44-90d7-ae4afd3322af-kube-api-access-zgv78" (OuterVolumeSpecName: "kube-api-access-zgv78") pod "5ae35e54-8396-4a44-90d7-ae4afd3322af" (UID: "5ae35e54-8396-4a44-90d7-ae4afd3322af"). InnerVolumeSpecName "kube-api-access-zgv78". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.528330 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e0a2-account-create-xss7d" event={"ID":"dc7dc25c-e0b6-4dc8-92f4-975dd74346a0","Type":"ContainerDied","Data":"767f2348215a12909125bab4024992f9be9ed81b2df539cb5b14d05f21908c76"} Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.528360 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="767f2348215a12909125bab4024992f9be9ed81b2df539cb5b14d05f21908c76" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.528443 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e0a2-account-create-xss7d" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.543894 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-eec8-account-create-snmtp" event={"ID":"b935dbba-ca03-419e-95de-f1061fbc0999","Type":"ContainerDied","Data":"a2c953af2c37256708ceaff813f7bf5d47356f7668188ddf42483aee0931adfc"} Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.543930 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2c953af2c37256708ceaff813f7bf5d47356f7668188ddf42483aee0931adfc" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.544009 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-eec8-account-create-snmtp" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.550905 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"89e12262-3506-49db-b83c-e9a53974ad5d","Type":"ContainerDied","Data":"d3c3b5c16d1ed60f1fdf53659af98f1126eb3740de0da4f9ea843419e7af7ba9"} Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.550901 4944 generic.go:334] "Generic (PLEG): container finished" podID="89e12262-3506-49db-b83c-e9a53974ad5d" containerID="d3c3b5c16d1ed60f1fdf53659af98f1126eb3740de0da4f9ea843419e7af7ba9" exitCode=0 Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.554758 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2d40-account-create-l7jv5" event={"ID":"5ae35e54-8396-4a44-90d7-ae4afd3322af","Type":"ContainerDied","Data":"abdee8457052c756ad740afc7cf6414bb45826ad5b64991fd5fd1b17614b4820"} Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.554789 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abdee8457052c756ad740afc7cf6414bb45826ad5b64991fd5fd1b17614b4820" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.554865 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2d40-account-create-l7jv5" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.559070 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-d8m52" event={"ID":"60b786e3-e5e4-4503-9887-9a3faef1d5c0","Type":"ContainerDied","Data":"a48a5b30be83ad437a2d8550494ed3e5b2ff94ac6eeb271658df43bee55c4bc5"} Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.559572 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a48a5b30be83ad437a2d8550494ed3e5b2ff94ac6eeb271658df43bee55c4bc5" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.559644 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-d8m52" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.563158 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cw4g5" event={"ID":"b92b104c-e53e-4246-b5d0-099005284506","Type":"ContainerDied","Data":"dd2f68edf9b5d85498f177bfea600649f2d1a5dd919f99b7ad318e69e1bf167f"} Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.563189 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd2f68edf9b5d85498f177bfea600649f2d1a5dd919f99b7ad318e69e1bf167f" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.563337 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cw4g5" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.567980 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.623617 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbfmx\" (UniqueName: \"kubernetes.io/projected/b935dbba-ca03-419e-95de-f1061fbc0999-kube-api-access-hbfmx\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.623969 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7mkg\" (UniqueName: \"kubernetes.io/projected/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-kube-api-access-d7mkg\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.623985 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.623999 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgv78\" (UniqueName: \"kubernetes.io/projected/5ae35e54-8396-4a44-90d7-ae4afd3322af-kube-api-access-zgv78\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.624901 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ae35e54-8396-4a44-90d7-ae4afd3322af-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.624927 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfb6w\" (UniqueName: \"kubernetes.io/projected/b92b104c-e53e-4246-b5d0-099005284506-kube-api-access-tfb6w\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.731683 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-httpd-run\") pod \"89e12262-3506-49db-b83c-e9a53974ad5d\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.731758 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-scripts\") pod \"89e12262-3506-49db-b83c-e9a53974ad5d\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.731800 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"89e12262-3506-49db-b83c-e9a53974ad5d\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.731837 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-combined-ca-bundle\") pod \"89e12262-3506-49db-b83c-e9a53974ad5d\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.731861 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-logs\") pod \"89e12262-3506-49db-b83c-e9a53974ad5d\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.732030 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vmwp\" (UniqueName: \"kubernetes.io/projected/89e12262-3506-49db-b83c-e9a53974ad5d-kube-api-access-4vmwp\") pod \"89e12262-3506-49db-b83c-e9a53974ad5d\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.732084 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-config-data\") pod \"89e12262-3506-49db-b83c-e9a53974ad5d\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.732126 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-internal-tls-certs\") pod \"89e12262-3506-49db-b83c-e9a53974ad5d\" (UID: \"89e12262-3506-49db-b83c-e9a53974ad5d\") " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.732457 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "89e12262-3506-49db-b83c-e9a53974ad5d" (UID: "89e12262-3506-49db-b83c-e9a53974ad5d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.732984 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-logs" (OuterVolumeSpecName: "logs") pod "89e12262-3506-49db-b83c-e9a53974ad5d" (UID: "89e12262-3506-49db-b83c-e9a53974ad5d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.740193 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-scripts" (OuterVolumeSpecName: "scripts") pod "89e12262-3506-49db-b83c-e9a53974ad5d" (UID: "89e12262-3506-49db-b83c-e9a53974ad5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.740453 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "89e12262-3506-49db-b83c-e9a53974ad5d" (UID: "89e12262-3506-49db-b83c-e9a53974ad5d"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.749131 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89e12262-3506-49db-b83c-e9a53974ad5d-kube-api-access-4vmwp" (OuterVolumeSpecName: "kube-api-access-4vmwp") pod "89e12262-3506-49db-b83c-e9a53974ad5d" (UID: "89e12262-3506-49db-b83c-e9a53974ad5d"). InnerVolumeSpecName "kube-api-access-4vmwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.780924 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89e12262-3506-49db-b83c-e9a53974ad5d" (UID: "89e12262-3506-49db-b83c-e9a53974ad5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.803941 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-config-data" (OuterVolumeSpecName: "config-data") pod "89e12262-3506-49db-b83c-e9a53974ad5d" (UID: "89e12262-3506-49db-b83c-e9a53974ad5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.804629 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "89e12262-3506-49db-b83c-e9a53974ad5d" (UID: "89e12262-3506-49db-b83c-e9a53974ad5d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.834003 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.834039 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vmwp\" (UniqueName: \"kubernetes.io/projected/89e12262-3506-49db-b83c-e9a53974ad5d-kube-api-access-4vmwp\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.834080 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.834090 4944 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.834098 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/89e12262-3506-49db-b83c-e9a53974ad5d-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.834106 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.834132 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.834140 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89e12262-3506-49db-b83c-e9a53974ad5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.909526 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.936106 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:09 crc kubenswrapper[4944]: I1124 09:12:09.950624 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:12:09 crc kubenswrapper[4944]: W1124 09:12:09.963063 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf94b9960_fa37_49f7_9fa7_0e01f855992b.slice/crio-7aed0246358a3d9b0b2e65a77b8c37b393752525ae5e977f142c01e137b22d55 WatchSource:0}: Error finding container 7aed0246358a3d9b0b2e65a77b8c37b393752525ae5e977f142c01e137b22d55: Status 404 returned error can't find the container with id 7aed0246358a3d9b0b2e65a77b8c37b393752525ae5e977f142c01e137b22d55 Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.294061 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1221a90c-370a-4103-bd3c-b3dfe670b446" path="/var/lib/kubelet/pods/1221a90c-370a-4103-bd3c-b3dfe670b446/volumes" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.577610 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f94b9960-fa37-49f7-9fa7-0e01f855992b","Type":"ContainerStarted","Data":"7aed0246358a3d9b0b2e65a77b8c37b393752525ae5e977f142c01e137b22d55"} Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.579514 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"89e12262-3506-49db-b83c-e9a53974ad5d","Type":"ContainerDied","Data":"006a23afeae96526189b1fd3964137c2676694fcbc430fbb0cd51846186df0d3"} Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.579545 4944 scope.go:117] "RemoveContainer" containerID="d3c3b5c16d1ed60f1fdf53659af98f1126eb3740de0da4f9ea843419e7af7ba9" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.579604 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.608646 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.617891 4944 scope.go:117] "RemoveContainer" containerID="8bda4061dc19bf95a403f1262f4d42e018ccabe08f270939e7945ba45cb194b9" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.619877 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.631560 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:12:10 crc kubenswrapper[4944]: E1124 09:12:10.631967 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b935dbba-ca03-419e-95de-f1061fbc0999" containerName="mariadb-account-create" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.631983 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b935dbba-ca03-419e-95de-f1061fbc0999" containerName="mariadb-account-create" Nov 24 09:12:10 crc kubenswrapper[4944]: E1124 09:12:10.632001 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89e12262-3506-49db-b83c-e9a53974ad5d" containerName="glance-log" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632009 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="89e12262-3506-49db-b83c-e9a53974ad5d" containerName="glance-log" Nov 24 09:12:10 crc kubenswrapper[4944]: E1124 09:12:10.632021 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae35e54-8396-4a44-90d7-ae4afd3322af" containerName="mariadb-account-create" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632027 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae35e54-8396-4a44-90d7-ae4afd3322af" containerName="mariadb-account-create" Nov 24 09:12:10 crc kubenswrapper[4944]: E1124 09:12:10.632038 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60b786e3-e5e4-4503-9887-9a3faef1d5c0" containerName="mariadb-database-create" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632058 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="60b786e3-e5e4-4503-9887-9a3faef1d5c0" containerName="mariadb-database-create" Nov 24 09:12:10 crc kubenswrapper[4944]: E1124 09:12:10.632080 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89e12262-3506-49db-b83c-e9a53974ad5d" containerName="glance-httpd" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632087 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="89e12262-3506-49db-b83c-e9a53974ad5d" containerName="glance-httpd" Nov 24 09:12:10 crc kubenswrapper[4944]: E1124 09:12:10.632098 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b92b104c-e53e-4246-b5d0-099005284506" containerName="mariadb-database-create" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632104 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b92b104c-e53e-4246-b5d0-099005284506" containerName="mariadb-database-create" Nov 24 09:12:10 crc kubenswrapper[4944]: E1124 09:12:10.632117 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7dc25c-e0b6-4dc8-92f4-975dd74346a0" containerName="mariadb-account-create" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632122 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7dc25c-e0b6-4dc8-92f4-975dd74346a0" containerName="mariadb-account-create" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632266 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="89e12262-3506-49db-b83c-e9a53974ad5d" containerName="glance-httpd" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632286 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b92b104c-e53e-4246-b5d0-099005284506" containerName="mariadb-database-create" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632297 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc7dc25c-e0b6-4dc8-92f4-975dd74346a0" containerName="mariadb-account-create" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632305 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="89e12262-3506-49db-b83c-e9a53974ad5d" containerName="glance-log" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632318 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b935dbba-ca03-419e-95de-f1061fbc0999" containerName="mariadb-account-create" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632328 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="60b786e3-e5e4-4503-9887-9a3faef1d5c0" containerName="mariadb-database-create" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.632339 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae35e54-8396-4a44-90d7-ae4afd3322af" containerName="mariadb-account-create" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.633734 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.648226 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.649237 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.657883 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.767580 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.767687 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.767729 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.767767 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.767798 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-logs\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.767820 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jnqv\" (UniqueName: \"kubernetes.io/projected/7209b8a6-5a97-4611-8183-84f1db25e2d1-kube-api-access-4jnqv\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.767854 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.767946 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.870243 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.870302 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.870335 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.870358 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-logs\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.870379 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jnqv\" (UniqueName: \"kubernetes.io/projected/7209b8a6-5a97-4611-8183-84f1db25e2d1-kube-api-access-4jnqv\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.870406 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.870847 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.870963 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.871013 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-logs\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.871961 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.872513 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.881592 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.881698 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.882346 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.882555 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.891929 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jnqv\" (UniqueName: \"kubernetes.io/projected/7209b8a6-5a97-4611-8183-84f1db25e2d1-kube-api-access-4jnqv\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:10 crc kubenswrapper[4944]: I1124 09:12:10.902771 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " pod="openstack/glance-default-internal-api-0" Nov 24 09:12:11 crc kubenswrapper[4944]: I1124 09:12:11.082720 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:11 crc kubenswrapper[4944]: I1124 09:12:11.591443 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f94b9960-fa37-49f7-9fa7-0e01f855992b","Type":"ContainerStarted","Data":"9a5211b2a8ad4ed0c93e5f29ec422b4395da812067c859ba887bf13db9c4afb6"} Nov 24 09:12:11 crc kubenswrapper[4944]: I1124 09:12:11.591806 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f94b9960-fa37-49f7-9fa7-0e01f855992b","Type":"ContainerStarted","Data":"95d6f710f056de96375b740c77544fefc875b760b049fb6331a5d9ea4510ffb4"} Nov 24 09:12:11 crc kubenswrapper[4944]: I1124 09:12:11.615168 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.6151483840000003 podStartE2EDuration="3.615148384s" podCreationTimestamp="2025-11-24 09:12:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:12:11.608738299 +0000 UTC m=+1192.143178751" watchObservedRunningTime="2025-11-24 09:12:11.615148384 +0000 UTC m=+1192.149588856" Nov 24 09:12:11 crc kubenswrapper[4944]: I1124 09:12:11.683930 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:12:11 crc kubenswrapper[4944]: W1124 09:12:11.693094 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7209b8a6_5a97_4611_8183_84f1db25e2d1.slice/crio-ddc83015cc841e6bff66ee193ae085c8026e3ab37954bbe93c399a4b0318c625 WatchSource:0}: Error finding container ddc83015cc841e6bff66ee193ae085c8026e3ab37954bbe93c399a4b0318c625: Status 404 returned error can't find the container with id ddc83015cc841e6bff66ee193ae085c8026e3ab37954bbe93c399a4b0318c625 Nov 24 09:12:12 crc kubenswrapper[4944]: I1124 09:12:12.292340 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89e12262-3506-49db-b83c-e9a53974ad5d" path="/var/lib/kubelet/pods/89e12262-3506-49db-b83c-e9a53974ad5d/volumes" Nov 24 09:12:12 crc kubenswrapper[4944]: I1124 09:12:12.600841 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7209b8a6-5a97-4611-8183-84f1db25e2d1","Type":"ContainerStarted","Data":"b48283673309355d5ff56fb2136836b9670187973eb44ca2cc22b5681057bccf"} Nov 24 09:12:12 crc kubenswrapper[4944]: I1124 09:12:12.600888 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7209b8a6-5a97-4611-8183-84f1db25e2d1","Type":"ContainerStarted","Data":"ddc83015cc841e6bff66ee193ae085c8026e3ab37954bbe93c399a4b0318c625"} Nov 24 09:12:13 crc kubenswrapper[4944]: I1124 09:12:13.611183 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7209b8a6-5a97-4611-8183-84f1db25e2d1","Type":"ContainerStarted","Data":"71ca2485422737d39ab2f15fc0ad24cd930f9cf353b7e908277d85b6a71d3f14"} Nov 24 09:12:13 crc kubenswrapper[4944]: I1124 09:12:13.627347 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.627328215 podStartE2EDuration="3.627328215s" podCreationTimestamp="2025-11-24 09:12:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:12:13.626778977 +0000 UTC m=+1194.161219429" watchObservedRunningTime="2025-11-24 09:12:13.627328215 +0000 UTC m=+1194.161768677" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.346613 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ttlw6"] Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.348577 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.352164 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.352621 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-rrl49" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.353804 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.369894 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ttlw6"] Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.463603 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-scripts\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.463904 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-config-data\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.464145 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.464230 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcvgc\" (UniqueName: \"kubernetes.io/projected/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-kube-api-access-bcvgc\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.566527 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-scripts\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.566623 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-config-data\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.566689 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.566729 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcvgc\" (UniqueName: \"kubernetes.io/projected/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-kube-api-access-bcvgc\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.573738 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-scripts\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.578828 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.579327 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-config-data\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.599479 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcvgc\" (UniqueName: \"kubernetes.io/projected/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-kube-api-access-bcvgc\") pod \"nova-cell0-conductor-db-sync-ttlw6\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:15 crc kubenswrapper[4944]: I1124 09:12:15.675021 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.034612 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.211339 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-run-httpd\") pod \"5cbd5e05-4fba-4490-be02-8cc6441181cd\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.211433 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-log-httpd\") pod \"5cbd5e05-4fba-4490-be02-8cc6441181cd\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.211566 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-combined-ca-bundle\") pod \"5cbd5e05-4fba-4490-be02-8cc6441181cd\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.211609 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlqwc\" (UniqueName: \"kubernetes.io/projected/5cbd5e05-4fba-4490-be02-8cc6441181cd-kube-api-access-xlqwc\") pod \"5cbd5e05-4fba-4490-be02-8cc6441181cd\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.211661 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-config-data\") pod \"5cbd5e05-4fba-4490-be02-8cc6441181cd\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.211848 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-scripts\") pod \"5cbd5e05-4fba-4490-be02-8cc6441181cd\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.211896 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-sg-core-conf-yaml\") pod \"5cbd5e05-4fba-4490-be02-8cc6441181cd\" (UID: \"5cbd5e05-4fba-4490-be02-8cc6441181cd\") " Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.211888 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5cbd5e05-4fba-4490-be02-8cc6441181cd" (UID: "5cbd5e05-4fba-4490-be02-8cc6441181cd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:12:16 crc kubenswrapper[4944]: W1124 09:12:16.211985 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d3ec0c0_b512_4bca_aa21_57c13f34b5dd.slice/crio-f300753075cbb932aa020156deafda4d6cd50e8f5088b7b4fff1e0ae6dd2a48a WatchSource:0}: Error finding container f300753075cbb932aa020156deafda4d6cd50e8f5088b7b4fff1e0ae6dd2a48a: Status 404 returned error can't find the container with id f300753075cbb932aa020156deafda4d6cd50e8f5088b7b4fff1e0ae6dd2a48a Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.212356 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5cbd5e05-4fba-4490-be02-8cc6441181cd" (UID: "5cbd5e05-4fba-4490-be02-8cc6441181cd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.212638 4944 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.212663 4944 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cbd5e05-4fba-4490-be02-8cc6441181cd-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.216899 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cbd5e05-4fba-4490-be02-8cc6441181cd-kube-api-access-xlqwc" (OuterVolumeSpecName: "kube-api-access-xlqwc") pod "5cbd5e05-4fba-4490-be02-8cc6441181cd" (UID: "5cbd5e05-4fba-4490-be02-8cc6441181cd"). InnerVolumeSpecName "kube-api-access-xlqwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.218318 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-scripts" (OuterVolumeSpecName: "scripts") pod "5cbd5e05-4fba-4490-be02-8cc6441181cd" (UID: "5cbd5e05-4fba-4490-be02-8cc6441181cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.223970 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ttlw6"] Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.248269 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5cbd5e05-4fba-4490-be02-8cc6441181cd" (UID: "5cbd5e05-4fba-4490-be02-8cc6441181cd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.298379 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cbd5e05-4fba-4490-be02-8cc6441181cd" (UID: "5cbd5e05-4fba-4490-be02-8cc6441181cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.313786 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.313812 4944 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.313822 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.313832 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlqwc\" (UniqueName: \"kubernetes.io/projected/5cbd5e05-4fba-4490-be02-8cc6441181cd-kube-api-access-xlqwc\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.316418 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-config-data" (OuterVolumeSpecName: "config-data") pod "5cbd5e05-4fba-4490-be02-8cc6441181cd" (UID: "5cbd5e05-4fba-4490-be02-8cc6441181cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.416767 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cbd5e05-4fba-4490-be02-8cc6441181cd-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.640377 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ttlw6" event={"ID":"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd","Type":"ContainerStarted","Data":"f300753075cbb932aa020156deafda4d6cd50e8f5088b7b4fff1e0ae6dd2a48a"} Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.644942 4944 generic.go:334] "Generic (PLEG): container finished" podID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerID="137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33" exitCode=0 Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.644991 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cbd5e05-4fba-4490-be02-8cc6441181cd","Type":"ContainerDied","Data":"137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33"} Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.645026 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cbd5e05-4fba-4490-be02-8cc6441181cd","Type":"ContainerDied","Data":"52450ce3d2ed476be62417930f7158ac726f2b669dae59693f4b7944ce6fb2fd"} Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.645069 4944 scope.go:117] "RemoveContainer" containerID="b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.645076 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.666001 4944 scope.go:117] "RemoveContainer" containerID="e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.679987 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.697267 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.698914 4944 scope.go:117] "RemoveContainer" containerID="2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.713473 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:16 crc kubenswrapper[4944]: E1124 09:12:16.713883 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="ceilometer-central-agent" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.713907 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="ceilometer-central-agent" Nov 24 09:12:16 crc kubenswrapper[4944]: E1124 09:12:16.713931 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="proxy-httpd" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.713939 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="proxy-httpd" Nov 24 09:12:16 crc kubenswrapper[4944]: E1124 09:12:16.713954 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="ceilometer-notification-agent" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.713961 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="ceilometer-notification-agent" Nov 24 09:12:16 crc kubenswrapper[4944]: E1124 09:12:16.713976 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="sg-core" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.713982 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="sg-core" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.714183 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="sg-core" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.714215 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="ceilometer-notification-agent" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.714223 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="ceilometer-central-agent" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.714236 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" containerName="proxy-httpd" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.715876 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.719041 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.720165 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.721351 4944 scope.go:117] "RemoveContainer" containerID="137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.726209 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.751861 4944 scope.go:117] "RemoveContainer" containerID="b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c" Nov 24 09:12:16 crc kubenswrapper[4944]: E1124 09:12:16.753178 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c\": container with ID starting with b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c not found: ID does not exist" containerID="b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.753208 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c"} err="failed to get container status \"b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c\": rpc error: code = NotFound desc = could not find container \"b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c\": container with ID starting with b1f31c29582c8223f6b2314f11cec180fff9c0dff6e7647f66470aef8af6636c not found: ID does not exist" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.753229 4944 scope.go:117] "RemoveContainer" containerID="e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a" Nov 24 09:12:16 crc kubenswrapper[4944]: E1124 09:12:16.753629 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a\": container with ID starting with e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a not found: ID does not exist" containerID="e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.753655 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a"} err="failed to get container status \"e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a\": rpc error: code = NotFound desc = could not find container \"e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a\": container with ID starting with e540aeaa4ee2d810a3195b170f45230cf6c70d8e47db67594c6632b1c91e783a not found: ID does not exist" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.753668 4944 scope.go:117] "RemoveContainer" containerID="2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549" Nov 24 09:12:16 crc kubenswrapper[4944]: E1124 09:12:16.753937 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549\": container with ID starting with 2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549 not found: ID does not exist" containerID="2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.754018 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549"} err="failed to get container status \"2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549\": rpc error: code = NotFound desc = could not find container \"2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549\": container with ID starting with 2b2cb7f24e26a068c7ccb1529a9940ebfe9c55c0caa924b971bf0e101599d549 not found: ID does not exist" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.754107 4944 scope.go:117] "RemoveContainer" containerID="137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33" Nov 24 09:12:16 crc kubenswrapper[4944]: E1124 09:12:16.754441 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33\": container with ID starting with 137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33 not found: ID does not exist" containerID="137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.754463 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33"} err="failed to get container status \"137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33\": rpc error: code = NotFound desc = could not find container \"137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33\": container with ID starting with 137634c3981cefad38ad6d559b3c2c17f77449e66337ada88b884795c40d4c33 not found: ID does not exist" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.823395 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.823452 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-scripts\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.823490 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-config-data\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.823548 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmsss\" (UniqueName: \"kubernetes.io/projected/f0d8a153-db11-460f-9529-e910b102f501-kube-api-access-gmsss\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.823983 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.824229 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-run-httpd\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.824348 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-log-httpd\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.925583 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-run-httpd\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.925830 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-log-httpd\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.925909 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.925984 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-scripts\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.926116 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-config-data\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.926261 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmsss\" (UniqueName: \"kubernetes.io/projected/f0d8a153-db11-460f-9529-e910b102f501-kube-api-access-gmsss\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.926153 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-log-httpd\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.926461 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.926804 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-run-httpd\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.931950 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.932956 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.948889 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmsss\" (UniqueName: \"kubernetes.io/projected/f0d8a153-db11-460f-9529-e910b102f501-kube-api-access-gmsss\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.948923 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-config-data\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:16 crc kubenswrapper[4944]: I1124 09:12:16.949952 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-scripts\") pod \"ceilometer-0\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " pod="openstack/ceilometer-0" Nov 24 09:12:17 crc kubenswrapper[4944]: I1124 09:12:17.048637 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:12:17 crc kubenswrapper[4944]: I1124 09:12:17.524264 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:17 crc kubenswrapper[4944]: W1124 09:12:17.530149 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0d8a153_db11_460f_9529_e910b102f501.slice/crio-636bccc311f1d9a128780d41e452d168b08495c8984e0ad30214c841f3cc39c2 WatchSource:0}: Error finding container 636bccc311f1d9a128780d41e452d168b08495c8984e0ad30214c841f3cc39c2: Status 404 returned error can't find the container with id 636bccc311f1d9a128780d41e452d168b08495c8984e0ad30214c841f3cc39c2 Nov 24 09:12:17 crc kubenswrapper[4944]: I1124 09:12:17.656858 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d8a153-db11-460f-9529-e910b102f501","Type":"ContainerStarted","Data":"636bccc311f1d9a128780d41e452d168b08495c8984e0ad30214c841f3cc39c2"} Nov 24 09:12:18 crc kubenswrapper[4944]: I1124 09:12:18.291318 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cbd5e05-4fba-4490-be02-8cc6441181cd" path="/var/lib/kubelet/pods/5cbd5e05-4fba-4490-be02-8cc6441181cd/volumes" Nov 24 09:12:18 crc kubenswrapper[4944]: I1124 09:12:18.668955 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d8a153-db11-460f-9529-e910b102f501","Type":"ContainerStarted","Data":"5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9"} Nov 24 09:12:19 crc kubenswrapper[4944]: I1124 09:12:19.258517 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 09:12:19 crc kubenswrapper[4944]: I1124 09:12:19.258860 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 09:12:19 crc kubenswrapper[4944]: I1124 09:12:19.297935 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 09:12:19 crc kubenswrapper[4944]: I1124 09:12:19.306719 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 09:12:19 crc kubenswrapper[4944]: I1124 09:12:19.682986 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d8a153-db11-460f-9529-e910b102f501","Type":"ContainerStarted","Data":"b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5"} Nov 24 09:12:19 crc kubenswrapper[4944]: I1124 09:12:19.683036 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 09:12:19 crc kubenswrapper[4944]: I1124 09:12:19.683217 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 09:12:21 crc kubenswrapper[4944]: I1124 09:12:21.083576 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:21 crc kubenswrapper[4944]: I1124 09:12:21.084026 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:21 crc kubenswrapper[4944]: I1124 09:12:21.117307 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:21 crc kubenswrapper[4944]: I1124 09:12:21.128235 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:21 crc kubenswrapper[4944]: I1124 09:12:21.702161 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 09:12:21 crc kubenswrapper[4944]: I1124 09:12:21.702205 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 09:12:21 crc kubenswrapper[4944]: I1124 09:12:21.705761 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:21 crc kubenswrapper[4944]: I1124 09:12:21.705808 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:21 crc kubenswrapper[4944]: I1124 09:12:21.854778 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 09:12:22 crc kubenswrapper[4944]: I1124 09:12:22.245142 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 09:12:23 crc kubenswrapper[4944]: I1124 09:12:23.724228 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 09:12:23 crc kubenswrapper[4944]: I1124 09:12:23.724521 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 09:12:23 crc kubenswrapper[4944]: I1124 09:12:23.958013 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:24 crc kubenswrapper[4944]: I1124 09:12:24.099557 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 09:12:25 crc kubenswrapper[4944]: I1124 09:12:25.748632 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ttlw6" event={"ID":"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd","Type":"ContainerStarted","Data":"da767b6a401b73c320d8ff35a4d71eba18485e8bfac8bc5be97a2e7215f5f07b"} Nov 24 09:12:25 crc kubenswrapper[4944]: I1124 09:12:25.751581 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d8a153-db11-460f-9529-e910b102f501","Type":"ContainerStarted","Data":"6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846"} Nov 24 09:12:26 crc kubenswrapper[4944]: I1124 09:12:26.764343 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d8a153-db11-460f-9529-e910b102f501","Type":"ContainerStarted","Data":"799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4"} Nov 24 09:12:26 crc kubenswrapper[4944]: I1124 09:12:26.798441 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-ttlw6" podStartSLOduration=2.9523177990000002 podStartE2EDuration="11.798402696s" podCreationTimestamp="2025-11-24 09:12:15 +0000 UTC" firstStartedPulling="2025-11-24 09:12:16.216206256 +0000 UTC m=+1196.750646718" lastFinishedPulling="2025-11-24 09:12:25.062291153 +0000 UTC m=+1205.596731615" observedRunningTime="2025-11-24 09:12:25.78036551 +0000 UTC m=+1206.314805992" watchObservedRunningTime="2025-11-24 09:12:26.798402696 +0000 UTC m=+1207.332843158" Nov 24 09:12:26 crc kubenswrapper[4944]: I1124 09:12:26.806302 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.908474741 podStartE2EDuration="10.806279417s" podCreationTimestamp="2025-11-24 09:12:16 +0000 UTC" firstStartedPulling="2025-11-24 09:12:17.533080127 +0000 UTC m=+1198.067520599" lastFinishedPulling="2025-11-24 09:12:26.430884813 +0000 UTC m=+1206.965325275" observedRunningTime="2025-11-24 09:12:26.794293555 +0000 UTC m=+1207.328734037" watchObservedRunningTime="2025-11-24 09:12:26.806279417 +0000 UTC m=+1207.340719879" Nov 24 09:12:27 crc kubenswrapper[4944]: I1124 09:12:27.773889 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 09:12:34 crc kubenswrapper[4944]: I1124 09:12:34.839376 4944 generic.go:334] "Generic (PLEG): container finished" podID="3d3ec0c0-b512-4bca-aa21-57c13f34b5dd" containerID="da767b6a401b73c320d8ff35a4d71eba18485e8bfac8bc5be97a2e7215f5f07b" exitCode=0 Nov 24 09:12:34 crc kubenswrapper[4944]: I1124 09:12:34.839450 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ttlw6" event={"ID":"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd","Type":"ContainerDied","Data":"da767b6a401b73c320d8ff35a4d71eba18485e8bfac8bc5be97a2e7215f5f07b"} Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.146472 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.226999 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcvgc\" (UniqueName: \"kubernetes.io/projected/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-kube-api-access-bcvgc\") pod \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.227200 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-combined-ca-bundle\") pod \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.227231 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-config-data\") pod \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.227375 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-scripts\") pod \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\" (UID: \"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd\") " Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.233088 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-kube-api-access-bcvgc" (OuterVolumeSpecName: "kube-api-access-bcvgc") pod "3d3ec0c0-b512-4bca-aa21-57c13f34b5dd" (UID: "3d3ec0c0-b512-4bca-aa21-57c13f34b5dd"). InnerVolumeSpecName "kube-api-access-bcvgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.233684 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-scripts" (OuterVolumeSpecName: "scripts") pod "3d3ec0c0-b512-4bca-aa21-57c13f34b5dd" (UID: "3d3ec0c0-b512-4bca-aa21-57c13f34b5dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.253215 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d3ec0c0-b512-4bca-aa21-57c13f34b5dd" (UID: "3d3ec0c0-b512-4bca-aa21-57c13f34b5dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.254303 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-config-data" (OuterVolumeSpecName: "config-data") pod "3d3ec0c0-b512-4bca-aa21-57c13f34b5dd" (UID: "3d3ec0c0-b512-4bca-aa21-57c13f34b5dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.329343 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.329383 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.329392 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.329401 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcvgc\" (UniqueName: \"kubernetes.io/projected/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd-kube-api-access-bcvgc\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.856942 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ttlw6" event={"ID":"3d3ec0c0-b512-4bca-aa21-57c13f34b5dd","Type":"ContainerDied","Data":"f300753075cbb932aa020156deafda4d6cd50e8f5088b7b4fff1e0ae6dd2a48a"} Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.857318 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f300753075cbb932aa020156deafda4d6cd50e8f5088b7b4fff1e0ae6dd2a48a" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.856980 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ttlw6" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.960639 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:12:36 crc kubenswrapper[4944]: E1124 09:12:36.961135 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d3ec0c0-b512-4bca-aa21-57c13f34b5dd" containerName="nova-cell0-conductor-db-sync" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.961156 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d3ec0c0-b512-4bca-aa21-57c13f34b5dd" containerName="nova-cell0-conductor-db-sync" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.961389 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d3ec0c0-b512-4bca-aa21-57c13f34b5dd" containerName="nova-cell0-conductor-db-sync" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.962162 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.964342 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.964774 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-rrl49" Nov 24 09:12:36 crc kubenswrapper[4944]: I1124 09:12:36.985108 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.041660 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.041727 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.041807 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67gww\" (UniqueName: \"kubernetes.io/projected/a8d71630-681e-4126-bb64-e8b03f83ce97-kube-api-access-67gww\") pod \"nova-cell0-conductor-0\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.143147 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67gww\" (UniqueName: \"kubernetes.io/projected/a8d71630-681e-4126-bb64-e8b03f83ce97-kube-api-access-67gww\") pod \"nova-cell0-conductor-0\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.143276 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.143317 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.148303 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.148676 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.159850 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67gww\" (UniqueName: \"kubernetes.io/projected/a8d71630-681e-4126-bb64-e8b03f83ce97-kube-api-access-67gww\") pod \"nova-cell0-conductor-0\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.282005 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.717927 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:12:37 crc kubenswrapper[4944]: I1124 09:12:37.865606 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8d71630-681e-4126-bb64-e8b03f83ce97","Type":"ContainerStarted","Data":"27eb10684971e552edf6b8cd884f0217f8df9fbd74270aa468d962957d555e20"} Nov 24 09:12:38 crc kubenswrapper[4944]: I1124 09:12:38.874527 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8d71630-681e-4126-bb64-e8b03f83ce97","Type":"ContainerStarted","Data":"2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef"} Nov 24 09:12:38 crc kubenswrapper[4944]: I1124 09:12:38.874834 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:38 crc kubenswrapper[4944]: I1124 09:12:38.896441 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.896421537 podStartE2EDuration="2.896421537s" podCreationTimestamp="2025-11-24 09:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:12:38.889190276 +0000 UTC m=+1219.423630738" watchObservedRunningTime="2025-11-24 09:12:38.896421537 +0000 UTC m=+1219.430861999" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.307834 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.791114 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-5w7xp"] Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.793022 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.794992 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.807454 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.817178 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5w7xp"] Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.840566 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f49v6\" (UniqueName: \"kubernetes.io/projected/8e780a24-2c73-45db-89c0-00a4ec351220-kube-api-access-f49v6\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.840673 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-scripts\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.840706 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.840944 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-config-data\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.948218 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-config-data\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.948295 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f49v6\" (UniqueName: \"kubernetes.io/projected/8e780a24-2c73-45db-89c0-00a4ec351220-kube-api-access-f49v6\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.948333 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-scripts\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.948354 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.959277 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.959400 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-scripts\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.960154 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-config-data\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.976491 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 09:12:42 crc kubenswrapper[4944]: I1124 09:12:42.978078 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:42.984700 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:42.985650 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f49v6\" (UniqueName: \"kubernetes.io/projected/8e780a24-2c73-45db-89c0-00a4ec351220-kube-api-access-f49v6\") pod \"nova-cell0-cell-mapping-5w7xp\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:42.988577 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.050068 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6mqm\" (UniqueName: \"kubernetes.io/projected/2d79e248-2a4a-41af-b906-94282ec872a9-kube-api-access-s6mqm\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.050165 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-config-data\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.050200 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.050255 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d79e248-2a4a-41af-b906-94282ec872a9-logs\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.105949 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.113946 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.117704 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.134468 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.138191 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.158109 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-config-data\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.158176 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.158226 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d79e248-2a4a-41af-b906-94282ec872a9-logs\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.158301 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6mqm\" (UniqueName: \"kubernetes.io/projected/2d79e248-2a4a-41af-b906-94282ec872a9-kube-api-access-s6mqm\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.160359 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d79e248-2a4a-41af-b906-94282ec872a9-logs\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.190094 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-config-data\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.190653 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.192990 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6mqm\" (UniqueName: \"kubernetes.io/projected/2d79e248-2a4a-41af-b906-94282ec872a9-kube-api-access-s6mqm\") pod \"nova-api-0\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.243125 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.244705 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.250578 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.250787 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.257607 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-8wv9x"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.259103 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.264149 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.264205 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lpbr\" (UniqueName: \"kubernetes.io/projected/852c6531-d5c6-47ae-932b-a5f7804330a0-kube-api-access-6lpbr\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.264226 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-svc\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.264253 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.264279 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-config-data\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.264297 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-config\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.264339 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnmnl\" (UniqueName: \"kubernetes.io/projected/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-kube-api-access-xnmnl\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.264375 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.264402 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852c6531-d5c6-47ae-932b-a5f7804330a0-logs\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.264420 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.293945 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-8wv9x"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.370416 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.371806 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.372105 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.374006 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.374755 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.374811 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnmnl\" (UniqueName: \"kubernetes.io/projected/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-kube-api-access-xnmnl\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.374886 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-config-data\") pod \"nova-scheduler-0\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " pod="openstack/nova-scheduler-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.374916 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.374974 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852c6531-d5c6-47ae-932b-a5f7804330a0-logs\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.374999 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.375024 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.375063 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2m7q\" (UniqueName: \"kubernetes.io/projected/a11c08f3-0a37-41c0-8541-65f163f673d1-kube-api-access-p2m7q\") pod \"nova-scheduler-0\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " pod="openstack/nova-scheduler-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.375127 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4db6t\" (UniqueName: \"kubernetes.io/projected/0e351668-faec-414a-83c4-d62382e055d3-kube-api-access-4db6t\") pod \"nova-cell1-novncproxy-0\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.375144 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.375205 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lpbr\" (UniqueName: \"kubernetes.io/projected/852c6531-d5c6-47ae-932b-a5f7804330a0-kube-api-access-6lpbr\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.375225 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-svc\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.375273 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.375293 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " pod="openstack/nova-scheduler-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.375370 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-config-data\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.375413 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-config\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.377178 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-config\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.378436 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.378745 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852c6531-d5c6-47ae-932b-a5f7804330a0-logs\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.379452 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.381379 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.383363 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-svc\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.415244 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.415668 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.415682 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-config-data\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.425135 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lpbr\" (UniqueName: \"kubernetes.io/projected/852c6531-d5c6-47ae-932b-a5f7804330a0-kube-api-access-6lpbr\") pod \"nova-metadata-0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.437414 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnmnl\" (UniqueName: \"kubernetes.io/projected/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-kube-api-access-xnmnl\") pod \"dnsmasq-dns-bccf8f775-8wv9x\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.438073 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.478834 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4db6t\" (UniqueName: \"kubernetes.io/projected/0e351668-faec-414a-83c4-d62382e055d3-kube-api-access-4db6t\") pod \"nova-cell1-novncproxy-0\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.478917 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " pod="openstack/nova-scheduler-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.478970 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.479012 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-config-data\") pod \"nova-scheduler-0\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " pod="openstack/nova-scheduler-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.479085 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.479108 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2m7q\" (UniqueName: \"kubernetes.io/projected/a11c08f3-0a37-41c0-8541-65f163f673d1-kube-api-access-p2m7q\") pod \"nova-scheduler-0\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " pod="openstack/nova-scheduler-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.487307 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.487570 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.489642 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-config-data\") pod \"nova-scheduler-0\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " pod="openstack/nova-scheduler-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.496515 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " pod="openstack/nova-scheduler-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.504772 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4db6t\" (UniqueName: \"kubernetes.io/projected/0e351668-faec-414a-83c4-d62382e055d3-kube-api-access-4db6t\") pod \"nova-cell1-novncproxy-0\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.505594 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2m7q\" (UniqueName: \"kubernetes.io/projected/a11c08f3-0a37-41c0-8541-65f163f673d1-kube-api-access-p2m7q\") pod \"nova-scheduler-0\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " pod="openstack/nova-scheduler-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.610119 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.652961 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.758580 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.768646 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5w7xp"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.911970 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-p2c5n"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.913656 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.916486 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.919176 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.930704 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-p2c5n"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.941965 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5w7xp" event={"ID":"8e780a24-2c73-45db-89c0-00a4ec351220","Type":"ContainerStarted","Data":"5dc30db809c06dd395bf30ccff9c4705be8204332acc9054012c12fb3998877d"} Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.962599 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.990849 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pvjl\" (UniqueName: \"kubernetes.io/projected/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-kube-api-access-6pvjl\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.990910 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.990965 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-config-data\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:43 crc kubenswrapper[4944]: I1124 09:12:43.991104 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-scripts\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:44 crc kubenswrapper[4944]: W1124 09:12:44.055850 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod852c6531_d5c6_47ae_932b_a5f7804330a0.slice/crio-f1622a7290571967e83f623f79dd432ef935469f11b0f3e3c72e66cc53019af7 WatchSource:0}: Error finding container f1622a7290571967e83f623f79dd432ef935469f11b0f3e3c72e66cc53019af7: Status 404 returned error can't find the container with id f1622a7290571967e83f623f79dd432ef935469f11b0f3e3c72e66cc53019af7 Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.065543 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.093091 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pvjl\" (UniqueName: \"kubernetes.io/projected/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-kube-api-access-6pvjl\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.093142 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.093188 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-config-data\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.093269 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-scripts\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.098402 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-config-data\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.100423 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-scripts\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.100841 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.109725 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pvjl\" (UniqueName: \"kubernetes.io/projected/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-kube-api-access-6pvjl\") pod \"nova-cell1-conductor-db-sync-p2c5n\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:44 crc kubenswrapper[4944]: W1124 09:12:44.211538 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda11c08f3_0a37_41c0_8541_65f163f673d1.slice/crio-9baaf7ce55844d520188a2042d5e2b8c6a00bfa2dd3c27f5c0f5e7e396277403 WatchSource:0}: Error finding container 9baaf7ce55844d520188a2042d5e2b8c6a00bfa2dd3c27f5c0f5e7e396277403: Status 404 returned error can't find the container with id 9baaf7ce55844d520188a2042d5e2b8c6a00bfa2dd3c27f5c0f5e7e396277403 Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.216774 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.264660 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.323243 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-8wv9x"] Nov 24 09:12:44 crc kubenswrapper[4944]: W1124 09:12:44.330103 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8b63713_ae4a_4cb3_98f7_0a51c7924ff0.slice/crio-a776c62271e6aa1c29d87621bd1053914dfdcfc4171f7ff0770ce52af58a1362 WatchSource:0}: Error finding container a776c62271e6aa1c29d87621bd1053914dfdcfc4171f7ff0770ce52af58a1362: Status 404 returned error can't find the container with id a776c62271e6aa1c29d87621bd1053914dfdcfc4171f7ff0770ce52af58a1362 Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.416962 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:12:44 crc kubenswrapper[4944]: W1124 09:12:44.434070 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e351668_faec_414a_83c4_d62382e055d3.slice/crio-dd790ddf3ca0535b6e6f9fb92c3ae0f56c2c480c16ff30bf0994ddf4b95062ed WatchSource:0}: Error finding container dd790ddf3ca0535b6e6f9fb92c3ae0f56c2c480c16ff30bf0994ddf4b95062ed: Status 404 returned error can't find the container with id dd790ddf3ca0535b6e6f9fb92c3ae0f56c2c480c16ff30bf0994ddf4b95062ed Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.762447 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-p2c5n"] Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.953747 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5w7xp" event={"ID":"8e780a24-2c73-45db-89c0-00a4ec351220","Type":"ContainerStarted","Data":"4e421ffe1b1f553a1d508dca1e10832265fc9517f18241ac1ae0a844024768cc"} Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.955784 4944 generic.go:334] "Generic (PLEG): container finished" podID="e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" containerID="f0b36825e86a4f2197ef7bde88dbfd675a16b202341b45e48fc5bf1000094669" exitCode=0 Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.955853 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" event={"ID":"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0","Type":"ContainerDied","Data":"f0b36825e86a4f2197ef7bde88dbfd675a16b202341b45e48fc5bf1000094669"} Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.955880 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" event={"ID":"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0","Type":"ContainerStarted","Data":"a776c62271e6aa1c29d87621bd1053914dfdcfc4171f7ff0770ce52af58a1362"} Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.960801 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"852c6531-d5c6-47ae-932b-a5f7804330a0","Type":"ContainerStarted","Data":"f1622a7290571967e83f623f79dd432ef935469f11b0f3e3c72e66cc53019af7"} Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.964819 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d79e248-2a4a-41af-b906-94282ec872a9","Type":"ContainerStarted","Data":"a623c290e3c93e5db76f254eb45501d17d9406c5056bafa6fa1c26f67179e94f"} Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.966833 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-p2c5n" event={"ID":"df085b2d-dbd0-4625-a905-a5f2c1dbedd8","Type":"ContainerStarted","Data":"b86407faaa0bba6a996005ade43cc750326f4ad0ba708600e9c924e3f039218f"} Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.972724 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a11c08f3-0a37-41c0-8541-65f163f673d1","Type":"ContainerStarted","Data":"9baaf7ce55844d520188a2042d5e2b8c6a00bfa2dd3c27f5c0f5e7e396277403"} Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.975434 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0e351668-faec-414a-83c4-d62382e055d3","Type":"ContainerStarted","Data":"dd790ddf3ca0535b6e6f9fb92c3ae0f56c2c480c16ff30bf0994ddf4b95062ed"} Nov 24 09:12:44 crc kubenswrapper[4944]: I1124 09:12:44.995244 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-5w7xp" podStartSLOduration=2.995127391 podStartE2EDuration="2.995127391s" podCreationTimestamp="2025-11-24 09:12:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:12:44.970465376 +0000 UTC m=+1225.504905838" watchObservedRunningTime="2025-11-24 09:12:44.995127391 +0000 UTC m=+1225.529567853" Nov 24 09:12:45 crc kubenswrapper[4944]: I1124 09:12:45.010180 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-p2c5n" podStartSLOduration=2.010155791 podStartE2EDuration="2.010155791s" podCreationTimestamp="2025-11-24 09:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:12:45.003134507 +0000 UTC m=+1225.537574979" watchObservedRunningTime="2025-11-24 09:12:45.010155791 +0000 UTC m=+1225.544596253" Nov 24 09:12:46 crc kubenswrapper[4944]: I1124 09:12:45.999929 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" event={"ID":"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0","Type":"ContainerStarted","Data":"577374ee469d4a9c77f3adeb7bc1d311545ffbc6d279cd8e1f90d2a17c173257"} Nov 24 09:12:46 crc kubenswrapper[4944]: I1124 09:12:46.001951 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:46 crc kubenswrapper[4944]: I1124 09:12:46.006082 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-p2c5n" event={"ID":"df085b2d-dbd0-4625-a905-a5f2c1dbedd8","Type":"ContainerStarted","Data":"afb36cdea5b1e686a6adc178f7749637e22fb3c8332fe0b88d9e35de15895874"} Nov 24 09:12:46 crc kubenswrapper[4944]: I1124 09:12:46.034013 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" podStartSLOduration=3.033987351 podStartE2EDuration="3.033987351s" podCreationTimestamp="2025-11-24 09:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:12:46.028528548 +0000 UTC m=+1226.562969010" watchObservedRunningTime="2025-11-24 09:12:46.033987351 +0000 UTC m=+1226.568427813" Nov 24 09:12:47 crc kubenswrapper[4944]: I1124 09:12:47.042114 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:47 crc kubenswrapper[4944]: I1124 09:12:47.057132 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:12:47 crc kubenswrapper[4944]: I1124 09:12:47.067768 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 09:12:48 crc kubenswrapper[4944]: I1124 09:12:48.031402 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d79e248-2a4a-41af-b906-94282ec872a9","Type":"ContainerStarted","Data":"296ec68d985a6e31a410b35534a786c564b8fe686fd119ca6a71a324d1ac92c6"} Nov 24 09:12:48 crc kubenswrapper[4944]: I1124 09:12:48.033778 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a11c08f3-0a37-41c0-8541-65f163f673d1","Type":"ContainerStarted","Data":"0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544"} Nov 24 09:12:48 crc kubenswrapper[4944]: I1124 09:12:48.037122 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0e351668-faec-414a-83c4-d62382e055d3","Type":"ContainerStarted","Data":"bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe"} Nov 24 09:12:48 crc kubenswrapper[4944]: I1124 09:12:48.037511 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0e351668-faec-414a-83c4-d62382e055d3" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe" gracePeriod=30 Nov 24 09:12:48 crc kubenswrapper[4944]: I1124 09:12:48.040697 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"852c6531-d5c6-47ae-932b-a5f7804330a0","Type":"ContainerStarted","Data":"0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da"} Nov 24 09:12:48 crc kubenswrapper[4944]: I1124 09:12:48.075766 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.881130358 podStartE2EDuration="5.075746996s" podCreationTimestamp="2025-11-24 09:12:43 +0000 UTC" firstStartedPulling="2025-11-24 09:12:44.216196646 +0000 UTC m=+1224.750637108" lastFinishedPulling="2025-11-24 09:12:47.410813284 +0000 UTC m=+1227.945253746" observedRunningTime="2025-11-24 09:12:48.051662558 +0000 UTC m=+1228.586103050" watchObservedRunningTime="2025-11-24 09:12:48.075746996 +0000 UTC m=+1228.610187458" Nov 24 09:12:48 crc kubenswrapper[4944]: I1124 09:12:48.076104 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.102878856 podStartE2EDuration="5.076040595s" podCreationTimestamp="2025-11-24 09:12:43 +0000 UTC" firstStartedPulling="2025-11-24 09:12:44.436206778 +0000 UTC m=+1224.970647240" lastFinishedPulling="2025-11-24 09:12:47.409368517 +0000 UTC m=+1227.943808979" observedRunningTime="2025-11-24 09:12:48.070413876 +0000 UTC m=+1228.604854338" watchObservedRunningTime="2025-11-24 09:12:48.076040595 +0000 UTC m=+1228.610481057" Nov 24 09:12:48 crc kubenswrapper[4944]: I1124 09:12:48.610651 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 09:12:48 crc kubenswrapper[4944]: I1124 09:12:48.759579 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.052020 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"852c6531-d5c6-47ae-932b-a5f7804330a0","Type":"ContainerStarted","Data":"6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949"} Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.052127 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="852c6531-d5c6-47ae-932b-a5f7804330a0" containerName="nova-metadata-log" containerID="cri-o://0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da" gracePeriod=30 Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.052217 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="852c6531-d5c6-47ae-932b-a5f7804330a0" containerName="nova-metadata-metadata" containerID="cri-o://6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949" gracePeriod=30 Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.058010 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d79e248-2a4a-41af-b906-94282ec872a9","Type":"ContainerStarted","Data":"9f233c38f7b12d120c06647ef8698552f28459f4dbab40420f4c7c1b64455333"} Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.089453 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.743554515 podStartE2EDuration="6.089426103s" podCreationTimestamp="2025-11-24 09:12:43 +0000 UTC" firstStartedPulling="2025-11-24 09:12:44.06224299 +0000 UTC m=+1224.596683452" lastFinishedPulling="2025-11-24 09:12:47.408114578 +0000 UTC m=+1227.942555040" observedRunningTime="2025-11-24 09:12:49.085604641 +0000 UTC m=+1229.620045103" watchObservedRunningTime="2025-11-24 09:12:49.089426103 +0000 UTC m=+1229.623866565" Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.743966 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.778143 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.309975498 podStartE2EDuration="7.777658978s" podCreationTimestamp="2025-11-24 09:12:42 +0000 UTC" firstStartedPulling="2025-11-24 09:12:43.941742979 +0000 UTC m=+1224.476183441" lastFinishedPulling="2025-11-24 09:12:47.409426459 +0000 UTC m=+1227.943866921" observedRunningTime="2025-11-24 09:12:49.136469642 +0000 UTC m=+1229.670910124" watchObservedRunningTime="2025-11-24 09:12:49.777658978 +0000 UTC m=+1230.312099430" Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.848149 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-config-data\") pod \"852c6531-d5c6-47ae-932b-a5f7804330a0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.848233 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lpbr\" (UniqueName: \"kubernetes.io/projected/852c6531-d5c6-47ae-932b-a5f7804330a0-kube-api-access-6lpbr\") pod \"852c6531-d5c6-47ae-932b-a5f7804330a0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.848362 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-combined-ca-bundle\") pod \"852c6531-d5c6-47ae-932b-a5f7804330a0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.848477 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852c6531-d5c6-47ae-932b-a5f7804330a0-logs\") pod \"852c6531-d5c6-47ae-932b-a5f7804330a0\" (UID: \"852c6531-d5c6-47ae-932b-a5f7804330a0\") " Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.850058 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/852c6531-d5c6-47ae-932b-a5f7804330a0-logs" (OuterVolumeSpecName: "logs") pod "852c6531-d5c6-47ae-932b-a5f7804330a0" (UID: "852c6531-d5c6-47ae-932b-a5f7804330a0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.860029 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/852c6531-d5c6-47ae-932b-a5f7804330a0-kube-api-access-6lpbr" (OuterVolumeSpecName: "kube-api-access-6lpbr") pod "852c6531-d5c6-47ae-932b-a5f7804330a0" (UID: "852c6531-d5c6-47ae-932b-a5f7804330a0"). InnerVolumeSpecName "kube-api-access-6lpbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.898578 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-config-data" (OuterVolumeSpecName: "config-data") pod "852c6531-d5c6-47ae-932b-a5f7804330a0" (UID: "852c6531-d5c6-47ae-932b-a5f7804330a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.913246 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "852c6531-d5c6-47ae-932b-a5f7804330a0" (UID: "852c6531-d5c6-47ae-932b-a5f7804330a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.950972 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.951006 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/852c6531-d5c6-47ae-932b-a5f7804330a0-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.951018 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/852c6531-d5c6-47ae-932b-a5f7804330a0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:49 crc kubenswrapper[4944]: I1124 09:12:49.951026 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lpbr\" (UniqueName: \"kubernetes.io/projected/852c6531-d5c6-47ae-932b-a5f7804330a0-kube-api-access-6lpbr\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.070397 4944 generic.go:334] "Generic (PLEG): container finished" podID="852c6531-d5c6-47ae-932b-a5f7804330a0" containerID="6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949" exitCode=0 Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.070440 4944 generic.go:334] "Generic (PLEG): container finished" podID="852c6531-d5c6-47ae-932b-a5f7804330a0" containerID="0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da" exitCode=143 Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.071582 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.073221 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"852c6531-d5c6-47ae-932b-a5f7804330a0","Type":"ContainerDied","Data":"6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949"} Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.073286 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"852c6531-d5c6-47ae-932b-a5f7804330a0","Type":"ContainerDied","Data":"0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da"} Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.073297 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"852c6531-d5c6-47ae-932b-a5f7804330a0","Type":"ContainerDied","Data":"f1622a7290571967e83f623f79dd432ef935469f11b0f3e3c72e66cc53019af7"} Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.073317 4944 scope.go:117] "RemoveContainer" containerID="6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.106383 4944 scope.go:117] "RemoveContainer" containerID="0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.116886 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.130439 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.136795 4944 scope.go:117] "RemoveContainer" containerID="6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949" Nov 24 09:12:50 crc kubenswrapper[4944]: E1124 09:12:50.137648 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949\": container with ID starting with 6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949 not found: ID does not exist" containerID="6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.137699 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949"} err="failed to get container status \"6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949\": rpc error: code = NotFound desc = could not find container \"6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949\": container with ID starting with 6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949 not found: ID does not exist" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.137730 4944 scope.go:117] "RemoveContainer" containerID="0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.140513 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:50 crc kubenswrapper[4944]: E1124 09:12:50.140991 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="852c6531-d5c6-47ae-932b-a5f7804330a0" containerName="nova-metadata-log" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.141016 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="852c6531-d5c6-47ae-932b-a5f7804330a0" containerName="nova-metadata-log" Nov 24 09:12:50 crc kubenswrapper[4944]: E1124 09:12:50.141032 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="852c6531-d5c6-47ae-932b-a5f7804330a0" containerName="nova-metadata-metadata" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.141040 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="852c6531-d5c6-47ae-932b-a5f7804330a0" containerName="nova-metadata-metadata" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.141290 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="852c6531-d5c6-47ae-932b-a5f7804330a0" containerName="nova-metadata-metadata" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.141318 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="852c6531-d5c6-47ae-932b-a5f7804330a0" containerName="nova-metadata-log" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.142665 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: E1124 09:12:50.142839 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da\": container with ID starting with 0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da not found: ID does not exist" containerID="0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.142879 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da"} err="failed to get container status \"0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da\": rpc error: code = NotFound desc = could not find container \"0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da\": container with ID starting with 0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da not found: ID does not exist" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.142909 4944 scope.go:117] "RemoveContainer" containerID="6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.147721 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.147738 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.151726 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949"} err="failed to get container status \"6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949\": rpc error: code = NotFound desc = could not find container \"6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949\": container with ID starting with 6f3f4c121ae9fe0cbfade3b8777e224f534a205d340d1927c2b2c5b8aa9fd949 not found: ID does not exist" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.151767 4944 scope.go:117] "RemoveContainer" containerID="0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.154756 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da"} err="failed to get container status \"0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da\": rpc error: code = NotFound desc = could not find container \"0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da\": container with ID starting with 0d0ac384465e653f6739b8fdd94be6f466ba46bed4db34581412e3069b2990da not found: ID does not exist" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.170522 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.256875 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.256978 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3da6b9a-0112-48fd-a267-eb5057bce602-logs\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.257170 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.257337 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-config-data\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.257582 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86nhx\" (UniqueName: \"kubernetes.io/projected/b3da6b9a-0112-48fd-a267-eb5057bce602-kube-api-access-86nhx\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.288149 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="852c6531-d5c6-47ae-932b-a5f7804330a0" path="/var/lib/kubelet/pods/852c6531-d5c6-47ae-932b-a5f7804330a0/volumes" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.359119 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3da6b9a-0112-48fd-a267-eb5057bce602-logs\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.359195 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.359252 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-config-data\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.359338 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86nhx\" (UniqueName: \"kubernetes.io/projected/b3da6b9a-0112-48fd-a267-eb5057bce602-kube-api-access-86nhx\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.359408 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.361466 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3da6b9a-0112-48fd-a267-eb5057bce602-logs\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.364028 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-config-data\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.364673 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.365627 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.380866 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86nhx\" (UniqueName: \"kubernetes.io/projected/b3da6b9a-0112-48fd-a267-eb5057bce602-kube-api-access-86nhx\") pod \"nova-metadata-0\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.469900 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:12:50 crc kubenswrapper[4944]: I1124 09:12:50.930887 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:50 crc kubenswrapper[4944]: W1124 09:12:50.941109 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3da6b9a_0112_48fd_a267_eb5057bce602.slice/crio-55be60bd1b922e8c963cfff3cf7615800b562a72547395e97f20997f96590fe0 WatchSource:0}: Error finding container 55be60bd1b922e8c963cfff3cf7615800b562a72547395e97f20997f96590fe0: Status 404 returned error can't find the container with id 55be60bd1b922e8c963cfff3cf7615800b562a72547395e97f20997f96590fe0 Nov 24 09:12:51 crc kubenswrapper[4944]: I1124 09:12:51.084301 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3da6b9a-0112-48fd-a267-eb5057bce602","Type":"ContainerStarted","Data":"55be60bd1b922e8c963cfff3cf7615800b562a72547395e97f20997f96590fe0"} Nov 24 09:12:51 crc kubenswrapper[4944]: I1124 09:12:51.318274 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:12:51 crc kubenswrapper[4944]: I1124 09:12:51.318761 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="ca22541f-1cf6-4ddf-9da8-23faad6ef99e" containerName="kube-state-metrics" containerID="cri-o://27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e" gracePeriod=30 Nov 24 09:12:51 crc kubenswrapper[4944]: I1124 09:12:51.848796 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 09:12:51 crc kubenswrapper[4944]: I1124 09:12:51.896841 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpxqs\" (UniqueName: \"kubernetes.io/projected/ca22541f-1cf6-4ddf-9da8-23faad6ef99e-kube-api-access-bpxqs\") pod \"ca22541f-1cf6-4ddf-9da8-23faad6ef99e\" (UID: \"ca22541f-1cf6-4ddf-9da8-23faad6ef99e\") " Nov 24 09:12:51 crc kubenswrapper[4944]: I1124 09:12:51.904263 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca22541f-1cf6-4ddf-9da8-23faad6ef99e-kube-api-access-bpxqs" (OuterVolumeSpecName: "kube-api-access-bpxqs") pod "ca22541f-1cf6-4ddf-9da8-23faad6ef99e" (UID: "ca22541f-1cf6-4ddf-9da8-23faad6ef99e"). InnerVolumeSpecName "kube-api-access-bpxqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.000994 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpxqs\" (UniqueName: \"kubernetes.io/projected/ca22541f-1cf6-4ddf-9da8-23faad6ef99e-kube-api-access-bpxqs\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.092700 4944 generic.go:334] "Generic (PLEG): container finished" podID="ca22541f-1cf6-4ddf-9da8-23faad6ef99e" containerID="27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e" exitCode=2 Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.092760 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ca22541f-1cf6-4ddf-9da8-23faad6ef99e","Type":"ContainerDied","Data":"27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e"} Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.092785 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ca22541f-1cf6-4ddf-9da8-23faad6ef99e","Type":"ContainerDied","Data":"6eaf15f7622eac875b82f0db2a8c6ddb4dee3321c2a7ba16ab3010bb51e08e02"} Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.092801 4944 scope.go:117] "RemoveContainer" containerID="27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.093170 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.103009 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3da6b9a-0112-48fd-a267-eb5057bce602","Type":"ContainerStarted","Data":"36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982"} Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.103075 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3da6b9a-0112-48fd-a267-eb5057bce602","Type":"ContainerStarted","Data":"f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c"} Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.123281 4944 scope.go:117] "RemoveContainer" containerID="27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e" Nov 24 09:12:52 crc kubenswrapper[4944]: E1124 09:12:52.127256 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e\": container with ID starting with 27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e not found: ID does not exist" containerID="27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.127308 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e"} err="failed to get container status \"27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e\": rpc error: code = NotFound desc = could not find container \"27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e\": container with ID starting with 27d6da9ff64e76caf34d973447f2ac145a71a81a27eec69dee13a90c8517026e not found: ID does not exist" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.137831 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.13780975 podStartE2EDuration="2.13780975s" podCreationTimestamp="2025-11-24 09:12:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:12:52.120676693 +0000 UTC m=+1232.655117165" watchObservedRunningTime="2025-11-24 09:12:52.13780975 +0000 UTC m=+1232.672250202" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.162948 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.172251 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.179672 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:12:52 crc kubenswrapper[4944]: E1124 09:12:52.180172 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca22541f-1cf6-4ddf-9da8-23faad6ef99e" containerName="kube-state-metrics" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.180188 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca22541f-1cf6-4ddf-9da8-23faad6ef99e" containerName="kube-state-metrics" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.180381 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca22541f-1cf6-4ddf-9da8-23faad6ef99e" containerName="kube-state-metrics" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.181550 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.185531 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.185744 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.188842 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.289789 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca22541f-1cf6-4ddf-9da8-23faad6ef99e" path="/var/lib/kubelet/pods/ca22541f-1cf6-4ddf-9da8-23faad6ef99e/volumes" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.308659 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf68g\" (UniqueName: \"kubernetes.io/projected/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-api-access-qf68g\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.308725 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.308823 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.308891 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.410382 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf68g\" (UniqueName: \"kubernetes.io/projected/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-api-access-qf68g\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.410442 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.410518 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.410544 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.414514 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.415479 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.415738 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.438288 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf68g\" (UniqueName: \"kubernetes.io/projected/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-api-access-qf68g\") pod \"kube-state-metrics-0\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " pod="openstack/kube-state-metrics-0" Nov 24 09:12:52 crc kubenswrapper[4944]: I1124 09:12:52.505617 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.026130 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:12:53 crc kubenswrapper[4944]: W1124 09:12:53.032160 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2e0b8db_ac91_432d_a87c_e741e441c33a.slice/crio-91a4015d43e34cead4d81b303d5fd64d6ce15ca8f651b251392b1550a7892360 WatchSource:0}: Error finding container 91a4015d43e34cead4d81b303d5fd64d6ce15ca8f651b251392b1550a7892360: Status 404 returned error can't find the container with id 91a4015d43e34cead4d81b303d5fd64d6ce15ca8f651b251392b1550a7892360 Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.113789 4944 generic.go:334] "Generic (PLEG): container finished" podID="8e780a24-2c73-45db-89c0-00a4ec351220" containerID="4e421ffe1b1f553a1d508dca1e10832265fc9517f18241ac1ae0a844024768cc" exitCode=0 Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.113870 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5w7xp" event={"ID":"8e780a24-2c73-45db-89c0-00a4ec351220","Type":"ContainerDied","Data":"4e421ffe1b1f553a1d508dca1e10832265fc9517f18241ac1ae0a844024768cc"} Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.116189 4944 generic.go:334] "Generic (PLEG): container finished" podID="df085b2d-dbd0-4625-a905-a5f2c1dbedd8" containerID="afb36cdea5b1e686a6adc178f7749637e22fb3c8332fe0b88d9e35de15895874" exitCode=0 Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.116264 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-p2c5n" event={"ID":"df085b2d-dbd0-4625-a905-a5f2c1dbedd8","Type":"ContainerDied","Data":"afb36cdea5b1e686a6adc178f7749637e22fb3c8332fe0b88d9e35de15895874"} Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.118879 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d2e0b8db-ac91-432d-a87c-e741e441c33a","Type":"ContainerStarted","Data":"91a4015d43e34cead4d81b303d5fd64d6ce15ca8f651b251392b1550a7892360"} Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.375319 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.375382 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.403854 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.404151 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="ceilometer-central-agent" containerID="cri-o://5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9" gracePeriod=30 Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.404266 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="proxy-httpd" containerID="cri-o://799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4" gracePeriod=30 Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.404307 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="sg-core" containerID="cri-o://6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846" gracePeriod=30 Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.404355 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="ceilometer-notification-agent" containerID="cri-o://b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5" gracePeriod=30 Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.611140 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.649844 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.656325 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.769963 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-qxz6c"] Nov 24 09:12:53 crc kubenswrapper[4944]: I1124 09:12:53.770371 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" podUID="9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" containerName="dnsmasq-dns" containerID="cri-o://bf0729a1ce61e434d095faba09fa37d27712f335d1509123cdcf54575a7ac2a4" gracePeriod=10 Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.135003 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d2e0b8db-ac91-432d-a87c-e741e441c33a","Type":"ContainerStarted","Data":"cb293c82a7b8cd484e5348b25efb88a1b9008ec1eeee27738a2aeda8a4a18930"} Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.136442 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.140949 4944 generic.go:334] "Generic (PLEG): container finished" podID="f0d8a153-db11-460f-9529-e910b102f501" containerID="799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4" exitCode=0 Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.140973 4944 generic.go:334] "Generic (PLEG): container finished" podID="f0d8a153-db11-460f-9529-e910b102f501" containerID="6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846" exitCode=2 Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.140983 4944 generic.go:334] "Generic (PLEG): container finished" podID="f0d8a153-db11-460f-9529-e910b102f501" containerID="5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9" exitCode=0 Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.141073 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d8a153-db11-460f-9529-e910b102f501","Type":"ContainerDied","Data":"799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4"} Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.141105 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d8a153-db11-460f-9529-e910b102f501","Type":"ContainerDied","Data":"6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846"} Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.141119 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d8a153-db11-460f-9529-e910b102f501","Type":"ContainerDied","Data":"5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9"} Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.144255 4944 generic.go:334] "Generic (PLEG): container finished" podID="9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" containerID="bf0729a1ce61e434d095faba09fa37d27712f335d1509123cdcf54575a7ac2a4" exitCode=0 Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.144368 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" event={"ID":"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50","Type":"ContainerDied","Data":"bf0729a1ce61e434d095faba09fa37d27712f335d1509123cdcf54575a7ac2a4"} Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.160988 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.795928005 podStartE2EDuration="2.160964451s" podCreationTimestamp="2025-11-24 09:12:52 +0000 UTC" firstStartedPulling="2025-11-24 09:12:53.034560564 +0000 UTC m=+1233.569001026" lastFinishedPulling="2025-11-24 09:12:53.39959701 +0000 UTC m=+1233.934037472" observedRunningTime="2025-11-24 09:12:54.154135741 +0000 UTC m=+1234.688576213" watchObservedRunningTime="2025-11-24 09:12:54.160964451 +0000 UTC m=+1234.695404913" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.190166 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.359457 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.453247 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-svc\") pod \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.453533 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-nb\") pod \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.453597 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-config\") pod \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.453658 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-sb\") pod \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.453726 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-swift-storage-0\") pod \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.454666 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f94gr\" (UniqueName: \"kubernetes.io/projected/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-kube-api-access-f94gr\") pod \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\" (UID: \"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.458991 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2d79e248-2a4a-41af-b906-94282ec872a9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.180:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.459158 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2d79e248-2a4a-41af-b906-94282ec872a9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.180:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.460220 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-kube-api-access-f94gr" (OuterVolumeSpecName: "kube-api-access-f94gr") pod "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" (UID: "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50"). InnerVolumeSpecName "kube-api-access-f94gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.516452 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" (UID: "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.519549 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" (UID: "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.539553 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" (UID: "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.542224 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-config" (OuterVolumeSpecName: "config") pod "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" (UID: "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.545160 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" (UID: "9faf9b28-723f-4a15-8f4c-a15bfc8e6c50"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.556840 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f94gr\" (UniqueName: \"kubernetes.io/projected/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-kube-api-access-f94gr\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.556874 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.556888 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.556900 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.556909 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.556919 4944 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.561716 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.579361 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.658470 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-config-data\") pod \"8e780a24-2c73-45db-89c0-00a4ec351220\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.658542 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-scripts\") pod \"8e780a24-2c73-45db-89c0-00a4ec351220\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.658582 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pvjl\" (UniqueName: \"kubernetes.io/projected/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-kube-api-access-6pvjl\") pod \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.658641 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-combined-ca-bundle\") pod \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.658696 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-combined-ca-bundle\") pod \"8e780a24-2c73-45db-89c0-00a4ec351220\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.658738 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f49v6\" (UniqueName: \"kubernetes.io/projected/8e780a24-2c73-45db-89c0-00a4ec351220-kube-api-access-f49v6\") pod \"8e780a24-2c73-45db-89c0-00a4ec351220\" (UID: \"8e780a24-2c73-45db-89c0-00a4ec351220\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.658783 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-config-data\") pod \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.658887 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-scripts\") pod \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\" (UID: \"df085b2d-dbd0-4625-a905-a5f2c1dbedd8\") " Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.662337 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-scripts" (OuterVolumeSpecName: "scripts") pod "8e780a24-2c73-45db-89c0-00a4ec351220" (UID: "8e780a24-2c73-45db-89c0-00a4ec351220"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.663497 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e780a24-2c73-45db-89c0-00a4ec351220-kube-api-access-f49v6" (OuterVolumeSpecName: "kube-api-access-f49v6") pod "8e780a24-2c73-45db-89c0-00a4ec351220" (UID: "8e780a24-2c73-45db-89c0-00a4ec351220"). InnerVolumeSpecName "kube-api-access-f49v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.663605 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-kube-api-access-6pvjl" (OuterVolumeSpecName: "kube-api-access-6pvjl") pod "df085b2d-dbd0-4625-a905-a5f2c1dbedd8" (UID: "df085b2d-dbd0-4625-a905-a5f2c1dbedd8"). InnerVolumeSpecName "kube-api-access-6pvjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.672097 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-scripts" (OuterVolumeSpecName: "scripts") pod "df085b2d-dbd0-4625-a905-a5f2c1dbedd8" (UID: "df085b2d-dbd0-4625-a905-a5f2c1dbedd8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.695520 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-config-data" (OuterVolumeSpecName: "config-data") pod "8e780a24-2c73-45db-89c0-00a4ec351220" (UID: "8e780a24-2c73-45db-89c0-00a4ec351220"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.701471 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-config-data" (OuterVolumeSpecName: "config-data") pod "df085b2d-dbd0-4625-a905-a5f2c1dbedd8" (UID: "df085b2d-dbd0-4625-a905-a5f2c1dbedd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.703901 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e780a24-2c73-45db-89c0-00a4ec351220" (UID: "8e780a24-2c73-45db-89c0-00a4ec351220"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.713883 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df085b2d-dbd0-4625-a905-a5f2c1dbedd8" (UID: "df085b2d-dbd0-4625-a905-a5f2c1dbedd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.761510 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.761873 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.761937 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.762000 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pvjl\" (UniqueName: \"kubernetes.io/projected/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-kube-api-access-6pvjl\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.762078 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.762137 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e780a24-2c73-45db-89c0-00a4ec351220-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.762189 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f49v6\" (UniqueName: \"kubernetes.io/projected/8e780a24-2c73-45db-89c0-00a4ec351220-kube-api-access-f49v6\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:54 crc kubenswrapper[4944]: I1124 09:12:54.762271 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df085b2d-dbd0-4625-a905-a5f2c1dbedd8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.154137 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5w7xp" event={"ID":"8e780a24-2c73-45db-89c0-00a4ec351220","Type":"ContainerDied","Data":"5dc30db809c06dd395bf30ccff9c4705be8204332acc9054012c12fb3998877d"} Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.154172 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dc30db809c06dd395bf30ccff9c4705be8204332acc9054012c12fb3998877d" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.154222 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5w7xp" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.160956 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-p2c5n" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.160977 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-p2c5n" event={"ID":"df085b2d-dbd0-4625-a905-a5f2c1dbedd8","Type":"ContainerDied","Data":"b86407faaa0bba6a996005ade43cc750326f4ad0ba708600e9c924e3f039218f"} Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.161401 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b86407faaa0bba6a996005ade43cc750326f4ad0ba708600e9c924e3f039218f" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.163858 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.163853 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-qxz6c" event={"ID":"9faf9b28-723f-4a15-8f4c-a15bfc8e6c50","Type":"ContainerDied","Data":"7e24ee81704d1fd5d02b8d2a46731a59de09fa79b74c108799ccf0e29f71703c"} Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.163904 4944 scope.go:117] "RemoveContainer" containerID="bf0729a1ce61e434d095faba09fa37d27712f335d1509123cdcf54575a7ac2a4" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.192930 4944 scope.go:117] "RemoveContainer" containerID="33286e6f5da914e557f15218ff1cd6235479db1db7c60eb8b559b8d60d6343b5" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.306205 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-qxz6c"] Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.316248 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-qxz6c"] Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.328105 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:12:55 crc kubenswrapper[4944]: E1124 09:12:55.328600 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e780a24-2c73-45db-89c0-00a4ec351220" containerName="nova-manage" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.328618 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e780a24-2c73-45db-89c0-00a4ec351220" containerName="nova-manage" Nov 24 09:12:55 crc kubenswrapper[4944]: E1124 09:12:55.328633 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" containerName="init" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.328639 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" containerName="init" Nov 24 09:12:55 crc kubenswrapper[4944]: E1124 09:12:55.328664 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" containerName="dnsmasq-dns" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.328671 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" containerName="dnsmasq-dns" Nov 24 09:12:55 crc kubenswrapper[4944]: E1124 09:12:55.328680 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df085b2d-dbd0-4625-a905-a5f2c1dbedd8" containerName="nova-cell1-conductor-db-sync" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.328686 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="df085b2d-dbd0-4625-a905-a5f2c1dbedd8" containerName="nova-cell1-conductor-db-sync" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.328874 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" containerName="dnsmasq-dns" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.328897 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="df085b2d-dbd0-4625-a905-a5f2c1dbedd8" containerName="nova-cell1-conductor-db-sync" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.328913 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e780a24-2c73-45db-89c0-00a4ec351220" containerName="nova-manage" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.329706 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.333356 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.336430 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.399703 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.399979 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.400127 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hdvk\" (UniqueName: \"kubernetes.io/projected/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-kube-api-access-5hdvk\") pod \"nova-cell1-conductor-0\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.417294 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.417545 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2d79e248-2a4a-41af-b906-94282ec872a9" containerName="nova-api-log" containerID="cri-o://296ec68d985a6e31a410b35534a786c564b8fe686fd119ca6a71a324d1ac92c6" gracePeriod=30 Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.417962 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2d79e248-2a4a-41af-b906-94282ec872a9" containerName="nova-api-api" containerID="cri-o://9f233c38f7b12d120c06647ef8698552f28459f4dbab40420f4c7c1b64455333" gracePeriod=30 Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.429751 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.457483 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.457788 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b3da6b9a-0112-48fd-a267-eb5057bce602" containerName="nova-metadata-log" containerID="cri-o://f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c" gracePeriod=30 Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.457934 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b3da6b9a-0112-48fd-a267-eb5057bce602" containerName="nova-metadata-metadata" containerID="cri-o://36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982" gracePeriod=30 Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.471412 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.471544 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.501837 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.501896 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.501967 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hdvk\" (UniqueName: \"kubernetes.io/projected/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-kube-api-access-5hdvk\") pod \"nova-cell1-conductor-0\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.508350 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.522226 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.523526 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hdvk\" (UniqueName: \"kubernetes.io/projected/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-kube-api-access-5hdvk\") pod \"nova-cell1-conductor-0\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:55 crc kubenswrapper[4944]: I1124 09:12:55.671231 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.080800 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.187239 4944 generic.go:334] "Generic (PLEG): container finished" podID="b3da6b9a-0112-48fd-a267-eb5057bce602" containerID="36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982" exitCode=0 Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.187277 4944 generic.go:334] "Generic (PLEG): container finished" podID="b3da6b9a-0112-48fd-a267-eb5057bce602" containerID="f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c" exitCode=143 Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.187310 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.187291 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3da6b9a-0112-48fd-a267-eb5057bce602","Type":"ContainerDied","Data":"36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982"} Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.187447 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3da6b9a-0112-48fd-a267-eb5057bce602","Type":"ContainerDied","Data":"f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c"} Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.187464 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3da6b9a-0112-48fd-a267-eb5057bce602","Type":"ContainerDied","Data":"55be60bd1b922e8c963cfff3cf7615800b562a72547395e97f20997f96590fe0"} Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.187483 4944 scope.go:117] "RemoveContainer" containerID="36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.192720 4944 generic.go:334] "Generic (PLEG): container finished" podID="2d79e248-2a4a-41af-b906-94282ec872a9" containerID="296ec68d985a6e31a410b35534a786c564b8fe686fd119ca6a71a324d1ac92c6" exitCode=143 Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.192810 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d79e248-2a4a-41af-b906-94282ec872a9","Type":"ContainerDied","Data":"296ec68d985a6e31a410b35534a786c564b8fe686fd119ca6a71a324d1ac92c6"} Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.193246 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a11c08f3-0a37-41c0-8541-65f163f673d1" containerName="nova-scheduler-scheduler" containerID="cri-o://0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544" gracePeriod=30 Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.228182 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3da6b9a-0112-48fd-a267-eb5057bce602-logs\") pod \"b3da6b9a-0112-48fd-a267-eb5057bce602\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.228252 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-config-data\") pod \"b3da6b9a-0112-48fd-a267-eb5057bce602\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.228325 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86nhx\" (UniqueName: \"kubernetes.io/projected/b3da6b9a-0112-48fd-a267-eb5057bce602-kube-api-access-86nhx\") pod \"b3da6b9a-0112-48fd-a267-eb5057bce602\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.228432 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-combined-ca-bundle\") pod \"b3da6b9a-0112-48fd-a267-eb5057bce602\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.228511 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-nova-metadata-tls-certs\") pod \"b3da6b9a-0112-48fd-a267-eb5057bce602\" (UID: \"b3da6b9a-0112-48fd-a267-eb5057bce602\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.228659 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3da6b9a-0112-48fd-a267-eb5057bce602-logs" (OuterVolumeSpecName: "logs") pod "b3da6b9a-0112-48fd-a267-eb5057bce602" (UID: "b3da6b9a-0112-48fd-a267-eb5057bce602"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.228924 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3da6b9a-0112-48fd-a267-eb5057bce602-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.235009 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3da6b9a-0112-48fd-a267-eb5057bce602-kube-api-access-86nhx" (OuterVolumeSpecName: "kube-api-access-86nhx") pod "b3da6b9a-0112-48fd-a267-eb5057bce602" (UID: "b3da6b9a-0112-48fd-a267-eb5057bce602"). InnerVolumeSpecName "kube-api-access-86nhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.249323 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.260232 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-config-data" (OuterVolumeSpecName: "config-data") pod "b3da6b9a-0112-48fd-a267-eb5057bce602" (UID: "b3da6b9a-0112-48fd-a267-eb5057bce602"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.277104 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3da6b9a-0112-48fd-a267-eb5057bce602" (UID: "b3da6b9a-0112-48fd-a267-eb5057bce602"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.289269 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9faf9b28-723f-4a15-8f4c-a15bfc8e6c50" path="/var/lib/kubelet/pods/9faf9b28-723f-4a15-8f4c-a15bfc8e6c50/volumes" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.313549 4944 scope.go:117] "RemoveContainer" containerID="f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.318538 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b3da6b9a-0112-48fd-a267-eb5057bce602" (UID: "b3da6b9a-0112-48fd-a267-eb5057bce602"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.333510 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.333537 4944 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.333548 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3da6b9a-0112-48fd-a267-eb5057bce602-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.333556 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86nhx\" (UniqueName: \"kubernetes.io/projected/b3da6b9a-0112-48fd-a267-eb5057bce602-kube-api-access-86nhx\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.345633 4944 scope.go:117] "RemoveContainer" containerID="36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982" Nov 24 09:12:56 crc kubenswrapper[4944]: E1124 09:12:56.346080 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982\": container with ID starting with 36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982 not found: ID does not exist" containerID="36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.346114 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982"} err="failed to get container status \"36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982\": rpc error: code = NotFound desc = could not find container \"36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982\": container with ID starting with 36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982 not found: ID does not exist" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.346142 4944 scope.go:117] "RemoveContainer" containerID="f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c" Nov 24 09:12:56 crc kubenswrapper[4944]: E1124 09:12:56.346549 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c\": container with ID starting with f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c not found: ID does not exist" containerID="f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.346574 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c"} err="failed to get container status \"f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c\": rpc error: code = NotFound desc = could not find container \"f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c\": container with ID starting with f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c not found: ID does not exist" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.346590 4944 scope.go:117] "RemoveContainer" containerID="36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.346920 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982"} err="failed to get container status \"36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982\": rpc error: code = NotFound desc = could not find container \"36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982\": container with ID starting with 36533985ab118c84fdba85e54f99c799840bb025455f6aeb2cb580d4c0c78982 not found: ID does not exist" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.346942 4944 scope.go:117] "RemoveContainer" containerID="f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.347245 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c"} err="failed to get container status \"f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c\": rpc error: code = NotFound desc = could not find container \"f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c\": container with ID starting with f2216e156e6e15b26a48f169686e5eab1c2dfeb0b3047d5659494034ed6b402c not found: ID does not exist" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.500338 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.565721 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.581824 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.590326 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:56 crc kubenswrapper[4944]: E1124 09:12:56.590808 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="ceilometer-central-agent" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.590829 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="ceilometer-central-agent" Nov 24 09:12:56 crc kubenswrapper[4944]: E1124 09:12:56.590847 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3da6b9a-0112-48fd-a267-eb5057bce602" containerName="nova-metadata-log" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.590856 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3da6b9a-0112-48fd-a267-eb5057bce602" containerName="nova-metadata-log" Nov 24 09:12:56 crc kubenswrapper[4944]: E1124 09:12:56.590865 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="sg-core" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.590873 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="sg-core" Nov 24 09:12:56 crc kubenswrapper[4944]: E1124 09:12:56.590892 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="ceilometer-notification-agent" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.590900 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="ceilometer-notification-agent" Nov 24 09:12:56 crc kubenswrapper[4944]: E1124 09:12:56.590928 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="proxy-httpd" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.590937 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="proxy-httpd" Nov 24 09:12:56 crc kubenswrapper[4944]: E1124 09:12:56.590955 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3da6b9a-0112-48fd-a267-eb5057bce602" containerName="nova-metadata-metadata" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.590962 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3da6b9a-0112-48fd-a267-eb5057bce602" containerName="nova-metadata-metadata" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.591196 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="sg-core" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.591213 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="ceilometer-central-agent" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.591261 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="ceilometer-notification-agent" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.591273 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3da6b9a-0112-48fd-a267-eb5057bce602" containerName="nova-metadata-log" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.591294 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d8a153-db11-460f-9529-e910b102f501" containerName="proxy-httpd" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.591304 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3da6b9a-0112-48fd-a267-eb5057bce602" containerName="nova-metadata-metadata" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.592507 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.598525 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.598525 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.600295 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.641145 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-log-httpd\") pod \"f0d8a153-db11-460f-9529-e910b102f501\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.641211 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-sg-core-conf-yaml\") pod \"f0d8a153-db11-460f-9529-e910b102f501\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.641275 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-config-data\") pod \"f0d8a153-db11-460f-9529-e910b102f501\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.641300 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-scripts\") pod \"f0d8a153-db11-460f-9529-e910b102f501\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.641338 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmsss\" (UniqueName: \"kubernetes.io/projected/f0d8a153-db11-460f-9529-e910b102f501-kube-api-access-gmsss\") pod \"f0d8a153-db11-460f-9529-e910b102f501\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.641487 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-run-httpd\") pod \"f0d8a153-db11-460f-9529-e910b102f501\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.641539 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-combined-ca-bundle\") pod \"f0d8a153-db11-460f-9529-e910b102f501\" (UID: \"f0d8a153-db11-460f-9529-e910b102f501\") " Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.641740 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f0d8a153-db11-460f-9529-e910b102f501" (UID: "f0d8a153-db11-460f-9529-e910b102f501"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.642493 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f0d8a153-db11-460f-9529-e910b102f501" (UID: "f0d8a153-db11-460f-9529-e910b102f501"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.642659 4944 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.642687 4944 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d8a153-db11-460f-9529-e910b102f501-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.646034 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-scripts" (OuterVolumeSpecName: "scripts") pod "f0d8a153-db11-460f-9529-e910b102f501" (UID: "f0d8a153-db11-460f-9529-e910b102f501"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.646752 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0d8a153-db11-460f-9529-e910b102f501-kube-api-access-gmsss" (OuterVolumeSpecName: "kube-api-access-gmsss") pod "f0d8a153-db11-460f-9529-e910b102f501" (UID: "f0d8a153-db11-460f-9529-e910b102f501"). InnerVolumeSpecName "kube-api-access-gmsss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.669855 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f0d8a153-db11-460f-9529-e910b102f501" (UID: "f0d8a153-db11-460f-9529-e910b102f501"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.744434 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bae34b6c-03f3-4c99-b7c5-400541549ed3-logs\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.744491 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.744522 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.744555 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvpmp\" (UniqueName: \"kubernetes.io/projected/bae34b6c-03f3-4c99-b7c5-400541549ed3-kube-api-access-vvpmp\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.744588 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-config-data\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.744735 4944 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.744749 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.744762 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmsss\" (UniqueName: \"kubernetes.io/projected/f0d8a153-db11-460f-9529-e910b102f501-kube-api-access-gmsss\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.745110 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0d8a153-db11-460f-9529-e910b102f501" (UID: "f0d8a153-db11-460f-9529-e910b102f501"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.760947 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-config-data" (OuterVolumeSpecName: "config-data") pod "f0d8a153-db11-460f-9529-e910b102f501" (UID: "f0d8a153-db11-460f-9529-e910b102f501"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.846686 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bae34b6c-03f3-4c99-b7c5-400541549ed3-logs\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.846746 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.846773 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.846800 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvpmp\" (UniqueName: \"kubernetes.io/projected/bae34b6c-03f3-4c99-b7c5-400541549ed3-kube-api-access-vvpmp\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.846828 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-config-data\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.847526 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.847577 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d8a153-db11-460f-9529-e910b102f501-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.847656 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bae34b6c-03f3-4c99-b7c5-400541549ed3-logs\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.850539 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.850907 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.853579 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-config-data\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.870541 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvpmp\" (UniqueName: \"kubernetes.io/projected/bae34b6c-03f3-4c99-b7c5-400541549ed3-kube-api-access-vvpmp\") pod \"nova-metadata-0\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " pod="openstack/nova-metadata-0" Nov 24 09:12:56 crc kubenswrapper[4944]: I1124 09:12:56.919016 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:12:57 crc kubenswrapper[4944]: W1124 09:12:57.212914 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbae34b6c_03f3_4c99_b7c5_400541549ed3.slice/crio-04cde4650ac0dd5c85edff384cee6a87a8526b6865fb44bd27474cd0d27b0b7b WatchSource:0}: Error finding container 04cde4650ac0dd5c85edff384cee6a87a8526b6865fb44bd27474cd0d27b0b7b: Status 404 returned error can't find the container with id 04cde4650ac0dd5c85edff384cee6a87a8526b6865fb44bd27474cd0d27b0b7b Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.213776 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.222306 4944 generic.go:334] "Generic (PLEG): container finished" podID="f0d8a153-db11-460f-9529-e910b102f501" containerID="b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5" exitCode=0 Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.222405 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d8a153-db11-460f-9529-e910b102f501","Type":"ContainerDied","Data":"b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5"} Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.222411 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.222441 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d8a153-db11-460f-9529-e910b102f501","Type":"ContainerDied","Data":"636bccc311f1d9a128780d41e452d168b08495c8984e0ad30214c841f3cc39c2"} Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.222461 4944 scope.go:117] "RemoveContainer" containerID="799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.231478 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32","Type":"ContainerStarted","Data":"83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b"} Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.231532 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32","Type":"ContainerStarted","Data":"042d6d603e1db607911a0cb7bc01d5667ef75d0890485349d0ab3b3db6a46948"} Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.231667 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.266883 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.266839154 podStartE2EDuration="2.266839154s" podCreationTimestamp="2025-11-24 09:12:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:12:57.24716935 +0000 UTC m=+1237.781609812" watchObservedRunningTime="2025-11-24 09:12:57.266839154 +0000 UTC m=+1237.801279626" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.270729 4944 scope.go:117] "RemoveContainer" containerID="6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.296313 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.309476 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.318591 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.320714 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.321580 4944 scope.go:117] "RemoveContainer" containerID="b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.327209 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.328036 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.329772 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.363172 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.407405 4944 scope.go:117] "RemoveContainer" containerID="5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.437974 4944 scope.go:117] "RemoveContainer" containerID="799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4" Nov 24 09:12:57 crc kubenswrapper[4944]: E1124 09:12:57.439597 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4\": container with ID starting with 799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4 not found: ID does not exist" containerID="799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.439650 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4"} err="failed to get container status \"799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4\": rpc error: code = NotFound desc = could not find container \"799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4\": container with ID starting with 799aa73dd0df0d982790cf9d6b6087db9e39612b95b0741f9b626b84cb9d44d4 not found: ID does not exist" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.439686 4944 scope.go:117] "RemoveContainer" containerID="6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846" Nov 24 09:12:57 crc kubenswrapper[4944]: E1124 09:12:57.441573 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846\": container with ID starting with 6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846 not found: ID does not exist" containerID="6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.441616 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846"} err="failed to get container status \"6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846\": rpc error: code = NotFound desc = could not find container \"6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846\": container with ID starting with 6a093794f3c332bf998fc6bb3695cdedab88f3bff065d297d44846fc29cc2846 not found: ID does not exist" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.441645 4944 scope.go:117] "RemoveContainer" containerID="b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5" Nov 24 09:12:57 crc kubenswrapper[4944]: E1124 09:12:57.442010 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5\": container with ID starting with b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5 not found: ID does not exist" containerID="b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.442084 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5"} err="failed to get container status \"b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5\": rpc error: code = NotFound desc = could not find container \"b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5\": container with ID starting with b39120c1fc094eeedb09f85a7d4c640723aee314d846e52203e97673923e59d5 not found: ID does not exist" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.442120 4944 scope.go:117] "RemoveContainer" containerID="5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9" Nov 24 09:12:57 crc kubenswrapper[4944]: E1124 09:12:57.442597 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9\": container with ID starting with 5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9 not found: ID does not exist" containerID="5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.442628 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9"} err="failed to get container status \"5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9\": rpc error: code = NotFound desc = could not find container \"5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9\": container with ID starting with 5246aac1a1bda0e1b74ad66c8917b8bb415ae12e69a1ddc56af85cab0e3bd7d9 not found: ID does not exist" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.483271 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.483325 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-scripts\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.483343 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-log-httpd\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.483358 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-run-httpd\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.483401 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-config-data\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.483454 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.483478 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.483513 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxz2m\" (UniqueName: \"kubernetes.io/projected/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-kube-api-access-qxz2m\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.586350 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.586491 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-scripts\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.586546 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-log-httpd\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.586582 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-run-httpd\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.586668 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-config-data\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.586767 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.586859 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.587408 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-log-httpd\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.587658 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-run-httpd\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.587870 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxz2m\" (UniqueName: \"kubernetes.io/projected/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-kube-api-access-qxz2m\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.590608 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-scripts\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.591277 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-config-data\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.591825 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.593201 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.594261 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.610118 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxz2m\" (UniqueName: \"kubernetes.io/projected/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-kube-api-access-qxz2m\") pod \"ceilometer-0\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " pod="openstack/ceilometer-0" Nov 24 09:12:57 crc kubenswrapper[4944]: I1124 09:12:57.711419 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:12:58 crc kubenswrapper[4944]: I1124 09:12:58.224162 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:12:58 crc kubenswrapper[4944]: I1124 09:12:58.246391 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bae34b6c-03f3-4c99-b7c5-400541549ed3","Type":"ContainerStarted","Data":"09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18"} Nov 24 09:12:58 crc kubenswrapper[4944]: I1124 09:12:58.246450 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bae34b6c-03f3-4c99-b7c5-400541549ed3","Type":"ContainerStarted","Data":"d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0"} Nov 24 09:12:58 crc kubenswrapper[4944]: I1124 09:12:58.246462 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bae34b6c-03f3-4c99-b7c5-400541549ed3","Type":"ContainerStarted","Data":"04cde4650ac0dd5c85edff384cee6a87a8526b6865fb44bd27474cd0d27b0b7b"} Nov 24 09:12:58 crc kubenswrapper[4944]: I1124 09:12:58.253296 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7c98c3a-fe7c-4c62-9f3c-579024c160e9","Type":"ContainerStarted","Data":"55f1ed16cc6c0b53aa3f0e4a6a9edc3cbdc98fdf731b7ced25100451c84e862f"} Nov 24 09:12:58 crc kubenswrapper[4944]: I1124 09:12:58.272803 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.272775849 podStartE2EDuration="2.272775849s" podCreationTimestamp="2025-11-24 09:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:12:58.266416033 +0000 UTC m=+1238.800856525" watchObservedRunningTime="2025-11-24 09:12:58.272775849 +0000 UTC m=+1238.807216311" Nov 24 09:12:58 crc kubenswrapper[4944]: I1124 09:12:58.290998 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3da6b9a-0112-48fd-a267-eb5057bce602" path="/var/lib/kubelet/pods/b3da6b9a-0112-48fd-a267-eb5057bce602/volumes" Nov 24 09:12:58 crc kubenswrapper[4944]: I1124 09:12:58.291722 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0d8a153-db11-460f-9529-e910b102f501" path="/var/lib/kubelet/pods/f0d8a153-db11-460f-9529-e910b102f501/volumes" Nov 24 09:12:58 crc kubenswrapper[4944]: E1124 09:12:58.620278 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 09:12:58 crc kubenswrapper[4944]: E1124 09:12:58.621693 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 09:12:58 crc kubenswrapper[4944]: E1124 09:12:58.625579 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 09:12:58 crc kubenswrapper[4944]: E1124 09:12:58.625686 4944 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a11c08f3-0a37-41c0-8541-65f163f673d1" containerName="nova-scheduler-scheduler" Nov 24 09:12:59 crc kubenswrapper[4944]: I1124 09:12:59.268590 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7c98c3a-fe7c-4c62-9f3c-579024c160e9","Type":"ContainerStarted","Data":"9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91"} Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.279953 4944 generic.go:334] "Generic (PLEG): container finished" podID="2d79e248-2a4a-41af-b906-94282ec872a9" containerID="9f233c38f7b12d120c06647ef8698552f28459f4dbab40420f4c7c1b64455333" exitCode=0 Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.287105 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.289135 4944 generic.go:334] "Generic (PLEG): container finished" podID="a11c08f3-0a37-41c0-8541-65f163f673d1" containerID="0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544" exitCode=0 Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.318883 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d79e248-2a4a-41af-b906-94282ec872a9","Type":"ContainerDied","Data":"9f233c38f7b12d120c06647ef8698552f28459f4dbab40420f4c7c1b64455333"} Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.318964 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7c98c3a-fe7c-4c62-9f3c-579024c160e9","Type":"ContainerStarted","Data":"989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038"} Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.318985 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a11c08f3-0a37-41c0-8541-65f163f673d1","Type":"ContainerDied","Data":"0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544"} Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.319464 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a11c08f3-0a37-41c0-8541-65f163f673d1","Type":"ContainerDied","Data":"9baaf7ce55844d520188a2042d5e2b8c6a00bfa2dd3c27f5c0f5e7e396277403"} Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.321505 4944 scope.go:117] "RemoveContainer" containerID="0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.343086 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.347905 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2m7q\" (UniqueName: \"kubernetes.io/projected/a11c08f3-0a37-41c0-8541-65f163f673d1-kube-api-access-p2m7q\") pod \"a11c08f3-0a37-41c0-8541-65f163f673d1\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.348107 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-combined-ca-bundle\") pod \"a11c08f3-0a37-41c0-8541-65f163f673d1\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.348247 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-config-data\") pod \"a11c08f3-0a37-41c0-8541-65f163f673d1\" (UID: \"a11c08f3-0a37-41c0-8541-65f163f673d1\") " Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.362992 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a11c08f3-0a37-41c0-8541-65f163f673d1-kube-api-access-p2m7q" (OuterVolumeSpecName: "kube-api-access-p2m7q") pod "a11c08f3-0a37-41c0-8541-65f163f673d1" (UID: "a11c08f3-0a37-41c0-8541-65f163f673d1"). InnerVolumeSpecName "kube-api-access-p2m7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.380610 4944 scope.go:117] "RemoveContainer" containerID="0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544" Nov 24 09:13:00 crc kubenswrapper[4944]: E1124 09:13:00.381736 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544\": container with ID starting with 0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544 not found: ID does not exist" containerID="0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.381851 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544"} err="failed to get container status \"0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544\": rpc error: code = NotFound desc = could not find container \"0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544\": container with ID starting with 0cb5d2a846a92f4f4fff0166d58d08281a0dee6e4fdb37576280061547215544 not found: ID does not exist" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.416423 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-config-data" (OuterVolumeSpecName: "config-data") pod "a11c08f3-0a37-41c0-8541-65f163f673d1" (UID: "a11c08f3-0a37-41c0-8541-65f163f673d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.431198 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a11c08f3-0a37-41c0-8541-65f163f673d1" (UID: "a11c08f3-0a37-41c0-8541-65f163f673d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.451968 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6mqm\" (UniqueName: \"kubernetes.io/projected/2d79e248-2a4a-41af-b906-94282ec872a9-kube-api-access-s6mqm\") pod \"2d79e248-2a4a-41af-b906-94282ec872a9\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.452247 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-config-data\") pod \"2d79e248-2a4a-41af-b906-94282ec872a9\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.452292 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d79e248-2a4a-41af-b906-94282ec872a9-logs\") pod \"2d79e248-2a4a-41af-b906-94282ec872a9\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.452316 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-combined-ca-bundle\") pod \"2d79e248-2a4a-41af-b906-94282ec872a9\" (UID: \"2d79e248-2a4a-41af-b906-94282ec872a9\") " Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.452685 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2m7q\" (UniqueName: \"kubernetes.io/projected/a11c08f3-0a37-41c0-8541-65f163f673d1-kube-api-access-p2m7q\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.452703 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.452713 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a11c08f3-0a37-41c0-8541-65f163f673d1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.452949 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d79e248-2a4a-41af-b906-94282ec872a9-logs" (OuterVolumeSpecName: "logs") pod "2d79e248-2a4a-41af-b906-94282ec872a9" (UID: "2d79e248-2a4a-41af-b906-94282ec872a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.455377 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d79e248-2a4a-41af-b906-94282ec872a9-kube-api-access-s6mqm" (OuterVolumeSpecName: "kube-api-access-s6mqm") pod "2d79e248-2a4a-41af-b906-94282ec872a9" (UID: "2d79e248-2a4a-41af-b906-94282ec872a9"). InnerVolumeSpecName "kube-api-access-s6mqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.482756 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d79e248-2a4a-41af-b906-94282ec872a9" (UID: "2d79e248-2a4a-41af-b906-94282ec872a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.483152 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-config-data" (OuterVolumeSpecName: "config-data") pod "2d79e248-2a4a-41af-b906-94282ec872a9" (UID: "2d79e248-2a4a-41af-b906-94282ec872a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.556237 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6mqm\" (UniqueName: \"kubernetes.io/projected/2d79e248-2a4a-41af-b906-94282ec872a9-kube-api-access-s6mqm\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.556296 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.556313 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d79e248-2a4a-41af-b906-94282ec872a9-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:00 crc kubenswrapper[4944]: I1124 09:13:00.556327 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d79e248-2a4a-41af-b906-94282ec872a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.301261 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d79e248-2a4a-41af-b906-94282ec872a9","Type":"ContainerDied","Data":"a623c290e3c93e5db76f254eb45501d17d9406c5056bafa6fa1c26f67179e94f"} Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.301550 4944 scope.go:117] "RemoveContainer" containerID="9f233c38f7b12d120c06647ef8698552f28459f4dbab40420f4c7c1b64455333" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.301732 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.306148 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7c98c3a-fe7c-4c62-9f3c-579024c160e9","Type":"ContainerStarted","Data":"a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27"} Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.308401 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.335956 4944 scope.go:117] "RemoveContainer" containerID="296ec68d985a6e31a410b35534a786c564b8fe686fd119ca6a71a324d1ac92c6" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.344995 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.352418 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.359766 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.368456 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.374785 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:01 crc kubenswrapper[4944]: E1124 09:13:01.375222 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d79e248-2a4a-41af-b906-94282ec872a9" containerName="nova-api-api" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.375243 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d79e248-2a4a-41af-b906-94282ec872a9" containerName="nova-api-api" Nov 24 09:13:01 crc kubenswrapper[4944]: E1124 09:13:01.375282 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d79e248-2a4a-41af-b906-94282ec872a9" containerName="nova-api-log" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.375290 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d79e248-2a4a-41af-b906-94282ec872a9" containerName="nova-api-log" Nov 24 09:13:01 crc kubenswrapper[4944]: E1124 09:13:01.375309 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a11c08f3-0a37-41c0-8541-65f163f673d1" containerName="nova-scheduler-scheduler" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.375317 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a11c08f3-0a37-41c0-8541-65f163f673d1" containerName="nova-scheduler-scheduler" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.375490 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d79e248-2a4a-41af-b906-94282ec872a9" containerName="nova-api-log" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.375506 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d79e248-2a4a-41af-b906-94282ec872a9" containerName="nova-api-api" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.375524 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a11c08f3-0a37-41c0-8541-65f163f673d1" containerName="nova-scheduler-scheduler" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.376519 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.378934 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.409739 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.429215 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.430569 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.432856 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.450593 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.474365 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vxzw\" (UniqueName: \"kubernetes.io/projected/c45f64f4-e570-4379-9037-d8cf5712ee20-kube-api-access-9vxzw\") pod \"nova-scheduler-0\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.474411 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-config-data\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.474447 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-logs\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.474503 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k6xv\" (UniqueName: \"kubernetes.io/projected/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-kube-api-access-2k6xv\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.474536 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.474561 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.474610 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-config-data\") pod \"nova-scheduler-0\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.575938 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-logs\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.575976 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k6xv\" (UniqueName: \"kubernetes.io/projected/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-kube-api-access-2k6xv\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.576011 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.576059 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.576115 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-config-data\") pod \"nova-scheduler-0\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.576188 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vxzw\" (UniqueName: \"kubernetes.io/projected/c45f64f4-e570-4379-9037-d8cf5712ee20-kube-api-access-9vxzw\") pod \"nova-scheduler-0\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.576209 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-config-data\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.576610 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-logs\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.580098 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.580613 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.588760 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-config-data\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.593370 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-config-data\") pod \"nova-scheduler-0\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.598606 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vxzw\" (UniqueName: \"kubernetes.io/projected/c45f64f4-e570-4379-9037-d8cf5712ee20-kube-api-access-9vxzw\") pod \"nova-scheduler-0\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.598695 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k6xv\" (UniqueName: \"kubernetes.io/projected/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-kube-api-access-2k6xv\") pod \"nova-api-0\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.737238 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.752147 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.921643 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 09:13:01 crc kubenswrapper[4944]: I1124 09:13:01.921696 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 09:13:02 crc kubenswrapper[4944]: I1124 09:13:02.212663 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:02 crc kubenswrapper[4944]: I1124 09:13:02.273650 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:13:02 crc kubenswrapper[4944]: I1124 09:13:02.296351 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d79e248-2a4a-41af-b906-94282ec872a9" path="/var/lib/kubelet/pods/2d79e248-2a4a-41af-b906-94282ec872a9/volumes" Nov 24 09:13:02 crc kubenswrapper[4944]: I1124 09:13:02.297039 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a11c08f3-0a37-41c0-8541-65f163f673d1" path="/var/lib/kubelet/pods/a11c08f3-0a37-41c0-8541-65f163f673d1/volumes" Nov 24 09:13:02 crc kubenswrapper[4944]: I1124 09:13:02.342344 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7c98c3a-fe7c-4c62-9f3c-579024c160e9","Type":"ContainerStarted","Data":"8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac"} Nov 24 09:13:02 crc kubenswrapper[4944]: I1124 09:13:02.342530 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 09:13:02 crc kubenswrapper[4944]: I1124 09:13:02.344604 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d","Type":"ContainerStarted","Data":"1945c2588d4d96e819961532b6772010535ca612e349c4235550f0abe52de050"} Nov 24 09:13:02 crc kubenswrapper[4944]: I1124 09:13:02.353134 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c45f64f4-e570-4379-9037-d8cf5712ee20","Type":"ContainerStarted","Data":"e068be7bb5221a001225249f3c776ec49f1e05e9907f6a105757adb096947caf"} Nov 24 09:13:02 crc kubenswrapper[4944]: I1124 09:13:02.366157 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.123380636 podStartE2EDuration="5.366142391s" podCreationTimestamp="2025-11-24 09:12:57 +0000 UTC" firstStartedPulling="2025-11-24 09:12:58.225767462 +0000 UTC m=+1238.760207914" lastFinishedPulling="2025-11-24 09:13:01.468529207 +0000 UTC m=+1242.002969669" observedRunningTime="2025-11-24 09:13:02.361395177 +0000 UTC m=+1242.895835639" watchObservedRunningTime="2025-11-24 09:13:02.366142391 +0000 UTC m=+1242.900582853" Nov 24 09:13:02 crc kubenswrapper[4944]: I1124 09:13:02.519212 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 09:13:03 crc kubenswrapper[4944]: I1124 09:13:03.364418 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c45f64f4-e570-4379-9037-d8cf5712ee20","Type":"ContainerStarted","Data":"6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d"} Nov 24 09:13:03 crc kubenswrapper[4944]: I1124 09:13:03.369313 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d","Type":"ContainerStarted","Data":"adddd6de0ddb62e8560cdb8b84db23943795114824943ccedd1ac770f5584855"} Nov 24 09:13:03 crc kubenswrapper[4944]: I1124 09:13:03.369355 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d","Type":"ContainerStarted","Data":"50011423f53ad8de354eaa8fde0cf29012508191dee3ac249c077a00dd7415d9"} Nov 24 09:13:03 crc kubenswrapper[4944]: I1124 09:13:03.386012 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.385993823 podStartE2EDuration="2.385993823s" podCreationTimestamp="2025-11-24 09:13:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:13:03.380798236 +0000 UTC m=+1243.915238708" watchObservedRunningTime="2025-11-24 09:13:03.385993823 +0000 UTC m=+1243.920434285" Nov 24 09:13:03 crc kubenswrapper[4944]: I1124 09:13:03.407153 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.407126475 podStartE2EDuration="2.407126475s" podCreationTimestamp="2025-11-24 09:13:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:13:03.397770813 +0000 UTC m=+1243.932211295" watchObservedRunningTime="2025-11-24 09:13:03.407126475 +0000 UTC m=+1243.941566937" Nov 24 09:13:05 crc kubenswrapper[4944]: I1124 09:13:05.711185 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 09:13:06 crc kubenswrapper[4944]: I1124 09:13:06.752968 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 09:13:06 crc kubenswrapper[4944]: I1124 09:13:06.920185 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 09:13:06 crc kubenswrapper[4944]: I1124 09:13:06.920238 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 09:13:07 crc kubenswrapper[4944]: I1124 09:13:07.933343 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 09:13:07 crc kubenswrapper[4944]: I1124 09:13:07.935220 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 09:13:11 crc kubenswrapper[4944]: I1124 09:13:11.738137 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 09:13:11 crc kubenswrapper[4944]: I1124 09:13:11.739367 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 09:13:11 crc kubenswrapper[4944]: I1124 09:13:11.752815 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 09:13:11 crc kubenswrapper[4944]: I1124 09:13:11.781164 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 09:13:12 crc kubenswrapper[4944]: I1124 09:13:12.494274 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 09:13:12 crc kubenswrapper[4944]: I1124 09:13:12.820224 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 09:13:12 crc kubenswrapper[4944]: I1124 09:13:12.820254 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 09:13:16 crc kubenswrapper[4944]: I1124 09:13:16.925155 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 09:13:16 crc kubenswrapper[4944]: I1124 09:13:16.933008 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 09:13:16 crc kubenswrapper[4944]: I1124 09:13:16.933133 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 09:13:17 crc kubenswrapper[4944]: I1124 09:13:17.512347 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.444357 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.516934 4944 generic.go:334] "Generic (PLEG): container finished" podID="0e351668-faec-414a-83c4-d62382e055d3" containerID="bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe" exitCode=137 Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.516992 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.517029 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0e351668-faec-414a-83c4-d62382e055d3","Type":"ContainerDied","Data":"bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe"} Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.517086 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0e351668-faec-414a-83c4-d62382e055d3","Type":"ContainerDied","Data":"dd790ddf3ca0535b6e6f9fb92c3ae0f56c2c480c16ff30bf0994ddf4b95062ed"} Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.517105 4944 scope.go:117] "RemoveContainer" containerID="bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.537779 4944 scope.go:117] "RemoveContainer" containerID="bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe" Nov 24 09:13:18 crc kubenswrapper[4944]: E1124 09:13:18.538262 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe\": container with ID starting with bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe not found: ID does not exist" containerID="bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.538295 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe"} err="failed to get container status \"bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe\": rpc error: code = NotFound desc = could not find container \"bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe\": container with ID starting with bc24fb57deb038b55122d9971c26ac3d34eaa62a16c2100ab5cc5085733bcafe not found: ID does not exist" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.583881 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-config-data\") pod \"0e351668-faec-414a-83c4-d62382e055d3\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.584115 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-combined-ca-bundle\") pod \"0e351668-faec-414a-83c4-d62382e055d3\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.584202 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4db6t\" (UniqueName: \"kubernetes.io/projected/0e351668-faec-414a-83c4-d62382e055d3-kube-api-access-4db6t\") pod \"0e351668-faec-414a-83c4-d62382e055d3\" (UID: \"0e351668-faec-414a-83c4-d62382e055d3\") " Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.590258 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e351668-faec-414a-83c4-d62382e055d3-kube-api-access-4db6t" (OuterVolumeSpecName: "kube-api-access-4db6t") pod "0e351668-faec-414a-83c4-d62382e055d3" (UID: "0e351668-faec-414a-83c4-d62382e055d3"). InnerVolumeSpecName "kube-api-access-4db6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.612423 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-config-data" (OuterVolumeSpecName: "config-data") pod "0e351668-faec-414a-83c4-d62382e055d3" (UID: "0e351668-faec-414a-83c4-d62382e055d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.615360 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e351668-faec-414a-83c4-d62382e055d3" (UID: "0e351668-faec-414a-83c4-d62382e055d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.687122 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4db6t\" (UniqueName: \"kubernetes.io/projected/0e351668-faec-414a-83c4-d62382e055d3-kube-api-access-4db6t\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.687398 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.687432 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e351668-faec-414a-83c4-d62382e055d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.859647 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.873831 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.880970 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:13:18 crc kubenswrapper[4944]: E1124 09:13:18.881426 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e351668-faec-414a-83c4-d62382e055d3" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.881442 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e351668-faec-414a-83c4-d62382e055d3" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.882226 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e351668-faec-414a-83c4-d62382e055d3" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.889222 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.894317 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.895930 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.896272 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 09:13:18 crc kubenswrapper[4944]: I1124 09:13:18.904503 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.009498 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.009911 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.009981 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7mrb\" (UniqueName: \"kubernetes.io/projected/709b6d2b-7e0f-4f60-9bd2-edacb409a727-kube-api-access-g7mrb\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.010053 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.010085 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.111964 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.112071 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.112116 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7mrb\" (UniqueName: \"kubernetes.io/projected/709b6d2b-7e0f-4f60-9bd2-edacb409a727-kube-api-access-g7mrb\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.112167 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.112195 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.117665 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.117763 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.118133 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.119737 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.129906 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7mrb\" (UniqueName: \"kubernetes.io/projected/709b6d2b-7e0f-4f60-9bd2-edacb409a727-kube-api-access-g7mrb\") pod \"nova-cell1-novncproxy-0\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.248416 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:19 crc kubenswrapper[4944]: I1124 09:13:19.690461 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:13:19 crc kubenswrapper[4944]: W1124 09:13:19.692760 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod709b6d2b_7e0f_4f60_9bd2_edacb409a727.slice/crio-1fa5d802adbb6046d2c8102e2c47bf4b2d3d25fb826929f651055f3fb472fb9e WatchSource:0}: Error finding container 1fa5d802adbb6046d2c8102e2c47bf4b2d3d25fb826929f651055f3fb472fb9e: Status 404 returned error can't find the container with id 1fa5d802adbb6046d2c8102e2c47bf4b2d3d25fb826929f651055f3fb472fb9e Nov 24 09:13:20 crc kubenswrapper[4944]: I1124 09:13:20.289953 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e351668-faec-414a-83c4-d62382e055d3" path="/var/lib/kubelet/pods/0e351668-faec-414a-83c4-d62382e055d3/volumes" Nov 24 09:13:20 crc kubenswrapper[4944]: I1124 09:13:20.541375 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"709b6d2b-7e0f-4f60-9bd2-edacb409a727","Type":"ContainerStarted","Data":"bc869a248132ca4279a0e2f68eec18b02bc447d39b21ce0e334fb974d4d68487"} Nov 24 09:13:20 crc kubenswrapper[4944]: I1124 09:13:20.541420 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"709b6d2b-7e0f-4f60-9bd2-edacb409a727","Type":"ContainerStarted","Data":"1fa5d802adbb6046d2c8102e2c47bf4b2d3d25fb826929f651055f3fb472fb9e"} Nov 24 09:13:20 crc kubenswrapper[4944]: I1124 09:13:20.564322 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.5643010779999997 podStartE2EDuration="2.564301078s" podCreationTimestamp="2025-11-24 09:13:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:13:20.557948624 +0000 UTC m=+1261.092389096" watchObservedRunningTime="2025-11-24 09:13:20.564301078 +0000 UTC m=+1261.098741540" Nov 24 09:13:21 crc kubenswrapper[4944]: I1124 09:13:21.740790 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 09:13:21 crc kubenswrapper[4944]: I1124 09:13:21.740865 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 09:13:21 crc kubenswrapper[4944]: I1124 09:13:21.741289 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 09:13:21 crc kubenswrapper[4944]: I1124 09:13:21.741314 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 09:13:21 crc kubenswrapper[4944]: I1124 09:13:21.743919 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 09:13:21 crc kubenswrapper[4944]: I1124 09:13:21.749557 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 09:13:21 crc kubenswrapper[4944]: I1124 09:13:21.914141 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rsh4b"] Nov 24 09:13:21 crc kubenswrapper[4944]: I1124 09:13:21.917398 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:21 crc kubenswrapper[4944]: I1124 09:13:21.978996 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rsh4b"] Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.069727 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.070031 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.070180 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grfb9\" (UniqueName: \"kubernetes.io/projected/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-kube-api-access-grfb9\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.070283 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.070432 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-config\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.070511 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.172017 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-config\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.172104 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.172202 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.172231 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.172271 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grfb9\" (UniqueName: \"kubernetes.io/projected/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-kube-api-access-grfb9\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.172313 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.173649 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.174130 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-config\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.174555 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.174793 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.175126 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.192401 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grfb9\" (UniqueName: \"kubernetes.io/projected/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-kube-api-access-grfb9\") pod \"dnsmasq-dns-cd5cbd7b9-rsh4b\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.266838 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:22 crc kubenswrapper[4944]: I1124 09:13:22.762994 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rsh4b"] Nov 24 09:13:23 crc kubenswrapper[4944]: I1124 09:13:23.567948 4944 generic.go:334] "Generic (PLEG): container finished" podID="5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" containerID="1eb23035ae5379ce721468ef5b81b6c60c5a92e85e392991180915600706763e" exitCode=0 Nov 24 09:13:23 crc kubenswrapper[4944]: I1124 09:13:23.568079 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" event={"ID":"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef","Type":"ContainerDied","Data":"1eb23035ae5379ce721468ef5b81b6c60c5a92e85e392991180915600706763e"} Nov 24 09:13:23 crc kubenswrapper[4944]: I1124 09:13:23.568366 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" event={"ID":"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef","Type":"ContainerStarted","Data":"155986593ce3bec854130e84ea3700c6931f55e508517cb83b1791e49d0596ed"} Nov 24 09:13:23 crc kubenswrapper[4944]: I1124 09:13:23.831167 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:13:23 crc kubenswrapper[4944]: I1124 09:13:23.831599 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="ceilometer-central-agent" containerID="cri-o://9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91" gracePeriod=30 Nov 24 09:13:23 crc kubenswrapper[4944]: I1124 09:13:23.831735 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="proxy-httpd" containerID="cri-o://8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac" gracePeriod=30 Nov 24 09:13:23 crc kubenswrapper[4944]: I1124 09:13:23.831772 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="sg-core" containerID="cri-o://a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27" gracePeriod=30 Nov 24 09:13:23 crc kubenswrapper[4944]: I1124 09:13:23.831802 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="ceilometer-notification-agent" containerID="cri-o://989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038" gracePeriod=30 Nov 24 09:13:23 crc kubenswrapper[4944]: I1124 09:13:23.841826 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.190:3000/\": EOF" Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.249341 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.379208 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.579919 4944 generic.go:334] "Generic (PLEG): container finished" podID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerID="8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac" exitCode=0 Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.579958 4944 generic.go:334] "Generic (PLEG): container finished" podID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerID="a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27" exitCode=2 Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.579971 4944 generic.go:334] "Generic (PLEG): container finished" podID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerID="9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91" exitCode=0 Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.579963 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7c98c3a-fe7c-4c62-9f3c-579024c160e9","Type":"ContainerDied","Data":"8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac"} Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.580014 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7c98c3a-fe7c-4c62-9f3c-579024c160e9","Type":"ContainerDied","Data":"a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27"} Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.580030 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7c98c3a-fe7c-4c62-9f3c-579024c160e9","Type":"ContainerDied","Data":"9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91"} Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.582498 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" event={"ID":"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef","Type":"ContainerStarted","Data":"823389b569a8fd3fda96971a913dc73c72a4f283c5b7ab6145f8fb0dd9b6cb9e"} Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.582648 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerName="nova-api-log" containerID="cri-o://50011423f53ad8de354eaa8fde0cf29012508191dee3ac249c077a00dd7415d9" gracePeriod=30 Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.582719 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerName="nova-api-api" containerID="cri-o://adddd6de0ddb62e8560cdb8b84db23943795114824943ccedd1ac770f5584855" gracePeriod=30 Nov 24 09:13:24 crc kubenswrapper[4944]: I1124 09:13:24.608670 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" podStartSLOduration=3.608653951 podStartE2EDuration="3.608653951s" podCreationTimestamp="2025-11-24 09:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:13:24.606544142 +0000 UTC m=+1265.140984624" watchObservedRunningTime="2025-11-24 09:13:24.608653951 +0000 UTC m=+1265.143094413" Nov 24 09:13:25 crc kubenswrapper[4944]: I1124 09:13:25.594809 4944 generic.go:334] "Generic (PLEG): container finished" podID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerID="50011423f53ad8de354eaa8fde0cf29012508191dee3ac249c077a00dd7415d9" exitCode=143 Nov 24 09:13:25 crc kubenswrapper[4944]: I1124 09:13:25.595971 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d","Type":"ContainerDied","Data":"50011423f53ad8de354eaa8fde0cf29012508191dee3ac249c077a00dd7415d9"} Nov 24 09:13:25 crc kubenswrapper[4944]: I1124 09:13:25.596001 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.263219 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.369847 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-ceilometer-tls-certs\") pod \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.369915 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-scripts\") pod \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.370015 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-config-data\") pod \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.370035 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-combined-ca-bundle\") pod \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.370088 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-log-httpd\") pod \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.370121 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxz2m\" (UniqueName: \"kubernetes.io/projected/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-kube-api-access-qxz2m\") pod \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.370211 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-run-httpd\") pod \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.370230 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-sg-core-conf-yaml\") pod \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\" (UID: \"f7c98c3a-fe7c-4c62-9f3c-579024c160e9\") " Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.370932 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f7c98c3a-fe7c-4c62-9f3c-579024c160e9" (UID: "f7c98c3a-fe7c-4c62-9f3c-579024c160e9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.371061 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f7c98c3a-fe7c-4c62-9f3c-579024c160e9" (UID: "f7c98c3a-fe7c-4c62-9f3c-579024c160e9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.375869 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-kube-api-access-qxz2m" (OuterVolumeSpecName: "kube-api-access-qxz2m") pod "f7c98c3a-fe7c-4c62-9f3c-579024c160e9" (UID: "f7c98c3a-fe7c-4c62-9f3c-579024c160e9"). InnerVolumeSpecName "kube-api-access-qxz2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.376315 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-scripts" (OuterVolumeSpecName: "scripts") pod "f7c98c3a-fe7c-4c62-9f3c-579024c160e9" (UID: "f7c98c3a-fe7c-4c62-9f3c-579024c160e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.411008 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f7c98c3a-fe7c-4c62-9f3c-579024c160e9" (UID: "f7c98c3a-fe7c-4c62-9f3c-579024c160e9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.428766 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f7c98c3a-fe7c-4c62-9f3c-579024c160e9" (UID: "f7c98c3a-fe7c-4c62-9f3c-579024c160e9"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.455085 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7c98c3a-fe7c-4c62-9f3c-579024c160e9" (UID: "f7c98c3a-fe7c-4c62-9f3c-579024c160e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.472588 4944 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.473346 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxz2m\" (UniqueName: \"kubernetes.io/projected/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-kube-api-access-qxz2m\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.473368 4944 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.473382 4944 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.473396 4944 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.473407 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.473419 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.474501 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-config-data" (OuterVolumeSpecName: "config-data") pod "f7c98c3a-fe7c-4c62-9f3c-579024c160e9" (UID: "f7c98c3a-fe7c-4c62-9f3c-579024c160e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.575911 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c98c3a-fe7c-4c62-9f3c-579024c160e9-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.607935 4944 generic.go:334] "Generic (PLEG): container finished" podID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerID="989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038" exitCode=0 Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.608035 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.608084 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7c98c3a-fe7c-4c62-9f3c-579024c160e9","Type":"ContainerDied","Data":"989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038"} Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.608152 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7c98c3a-fe7c-4c62-9f3c-579024c160e9","Type":"ContainerDied","Data":"55f1ed16cc6c0b53aa3f0e4a6a9edc3cbdc98fdf731b7ced25100451c84e862f"} Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.608180 4944 scope.go:117] "RemoveContainer" containerID="8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.633286 4944 scope.go:117] "RemoveContainer" containerID="a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.645399 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.656115 4944 scope.go:117] "RemoveContainer" containerID="989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.664945 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.676175 4944 scope.go:117] "RemoveContainer" containerID="9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.676935 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:13:26 crc kubenswrapper[4944]: E1124 09:13:26.677487 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="proxy-httpd" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.677503 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="proxy-httpd" Nov 24 09:13:26 crc kubenswrapper[4944]: E1124 09:13:26.677516 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="ceilometer-notification-agent" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.677523 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="ceilometer-notification-agent" Nov 24 09:13:26 crc kubenswrapper[4944]: E1124 09:13:26.677541 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="sg-core" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.677548 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="sg-core" Nov 24 09:13:26 crc kubenswrapper[4944]: E1124 09:13:26.677566 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="ceilometer-central-agent" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.677573 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="ceilometer-central-agent" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.677811 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="proxy-httpd" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.677835 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="sg-core" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.677847 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="ceilometer-central-agent" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.677854 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" containerName="ceilometer-notification-agent" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.679652 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.682301 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.682836 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.682991 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.690249 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.761019 4944 scope.go:117] "RemoveContainer" containerID="8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac" Nov 24 09:13:26 crc kubenswrapper[4944]: E1124 09:13:26.761830 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac\": container with ID starting with 8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac not found: ID does not exist" containerID="8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.761878 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac"} err="failed to get container status \"8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac\": rpc error: code = NotFound desc = could not find container \"8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac\": container with ID starting with 8406ec42c52a3fd84adce376da758fc8206575fe03477ea8294c73e1f73209ac not found: ID does not exist" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.761910 4944 scope.go:117] "RemoveContainer" containerID="a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27" Nov 24 09:13:26 crc kubenswrapper[4944]: E1124 09:13:26.762383 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27\": container with ID starting with a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27 not found: ID does not exist" containerID="a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.762403 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27"} err="failed to get container status \"a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27\": rpc error: code = NotFound desc = could not find container \"a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27\": container with ID starting with a6a6cf80b5dc668d2ea876d73d13f2af82c4ef1d2d54639ffcf164983065aa27 not found: ID does not exist" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.762414 4944 scope.go:117] "RemoveContainer" containerID="989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038" Nov 24 09:13:26 crc kubenswrapper[4944]: E1124 09:13:26.762621 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038\": container with ID starting with 989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038 not found: ID does not exist" containerID="989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.762664 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038"} err="failed to get container status \"989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038\": rpc error: code = NotFound desc = could not find container \"989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038\": container with ID starting with 989ab9a17c031e240244425c335435cb64456a458fe610a043445b3e02651038 not found: ID does not exist" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.762677 4944 scope.go:117] "RemoveContainer" containerID="9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91" Nov 24 09:13:26 crc kubenswrapper[4944]: E1124 09:13:26.762987 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91\": container with ID starting with 9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91 not found: ID does not exist" containerID="9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.763012 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91"} err="failed to get container status \"9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91\": rpc error: code = NotFound desc = could not find container \"9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91\": container with ID starting with 9a5ea72aef257c8510e187fe5fdc6fa8816536872353a9e40d56c48b21d3bf91 not found: ID does not exist" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.882347 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-config-data\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.882782 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-log-httpd\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.883299 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.883365 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhtzl\" (UniqueName: \"kubernetes.io/projected/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-kube-api-access-qhtzl\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.883423 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.883454 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-run-httpd\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.883498 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.883538 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-scripts\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.987279 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.987338 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhtzl\" (UniqueName: \"kubernetes.io/projected/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-kube-api-access-qhtzl\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.987398 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.987418 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-run-httpd\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.987442 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.987481 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-scripts\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.987536 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-config-data\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.987557 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-log-httpd\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.988182 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-log-httpd\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.989212 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-run-httpd\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.993721 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.994860 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-config-data\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.995453 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-scripts\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:26 crc kubenswrapper[4944]: I1124 09:13:26.995968 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:27 crc kubenswrapper[4944]: I1124 09:13:27.001086 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:27 crc kubenswrapper[4944]: I1124 09:13:27.013394 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhtzl\" (UniqueName: \"kubernetes.io/projected/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-kube-api-access-qhtzl\") pod \"ceilometer-0\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " pod="openstack/ceilometer-0" Nov 24 09:13:27 crc kubenswrapper[4944]: I1124 09:13:27.042665 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:13:27 crc kubenswrapper[4944]: I1124 09:13:27.468388 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:13:27 crc kubenswrapper[4944]: W1124 09:13:27.474835 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cb23d74_ac96_4ce2_8a3b_0fe9aea10fe4.slice/crio-10020383316a643a77b2528f630a5557d012ead2d5b2ee865f54416fbdcbc49c WatchSource:0}: Error finding container 10020383316a643a77b2528f630a5557d012ead2d5b2ee865f54416fbdcbc49c: Status 404 returned error can't find the container with id 10020383316a643a77b2528f630a5557d012ead2d5b2ee865f54416fbdcbc49c Nov 24 09:13:27 crc kubenswrapper[4944]: I1124 09:13:27.620163 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4","Type":"ContainerStarted","Data":"10020383316a643a77b2528f630a5557d012ead2d5b2ee865f54416fbdcbc49c"} Nov 24 09:13:28 crc kubenswrapper[4944]: I1124 09:13:28.655551 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7c98c3a-fe7c-4c62-9f3c-579024c160e9" path="/var/lib/kubelet/pods/f7c98c3a-fe7c-4c62-9f3c-579024c160e9/volumes" Nov 24 09:13:28 crc kubenswrapper[4944]: I1124 09:13:28.681163 4944 generic.go:334] "Generic (PLEG): container finished" podID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerID="adddd6de0ddb62e8560cdb8b84db23943795114824943ccedd1ac770f5584855" exitCode=0 Nov 24 09:13:28 crc kubenswrapper[4944]: I1124 09:13:28.681211 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d","Type":"ContainerDied","Data":"adddd6de0ddb62e8560cdb8b84db23943795114824943ccedd1ac770f5584855"} Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.045913 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.218167 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-combined-ca-bundle\") pod \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.218849 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-config-data\") pod \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.218921 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-logs\") pod \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.219226 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k6xv\" (UniqueName: \"kubernetes.io/projected/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-kube-api-access-2k6xv\") pod \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\" (UID: \"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d\") " Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.235588 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-logs" (OuterVolumeSpecName: "logs") pod "23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" (UID: "23ad5a1c-7ea6-4909-a783-b14b80a4bb7d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.242924 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-kube-api-access-2k6xv" (OuterVolumeSpecName: "kube-api-access-2k6xv") pod "23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" (UID: "23ad5a1c-7ea6-4909-a783-b14b80a4bb7d"). InnerVolumeSpecName "kube-api-access-2k6xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.249822 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.262644 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-config-data" (OuterVolumeSpecName: "config-data") pod "23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" (UID: "23ad5a1c-7ea6-4909-a783-b14b80a4bb7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.265371 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" (UID: "23ad5a1c-7ea6-4909-a783-b14b80a4bb7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.289126 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.323065 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.323127 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.323143 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.323157 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k6xv\" (UniqueName: \"kubernetes.io/projected/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d-kube-api-access-2k6xv\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.696699 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23ad5a1c-7ea6-4909-a783-b14b80a4bb7d","Type":"ContainerDied","Data":"1945c2588d4d96e819961532b6772010535ca612e349c4235550f0abe52de050"} Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.696769 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.697111 4944 scope.go:117] "RemoveContainer" containerID="adddd6de0ddb62e8560cdb8b84db23943795114824943ccedd1ac770f5584855" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.701558 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4","Type":"ContainerStarted","Data":"f1c621c73eb738098cb49670f9fd9711e1d1a037fa44d09a782be7678f0546b3"} Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.701608 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4","Type":"ContainerStarted","Data":"73095b06c9c5460d64ea3a6e6ad293c16acb97dedadc392b482d5b45673686f5"} Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.720541 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.737682 4944 scope.go:117] "RemoveContainer" containerID="50011423f53ad8de354eaa8fde0cf29012508191dee3ac249c077a00dd7415d9" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.755992 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.780256 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.816174 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:29 crc kubenswrapper[4944]: E1124 09:13:29.816867 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerName="nova-api-api" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.816957 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerName="nova-api-api" Nov 24 09:13:29 crc kubenswrapper[4944]: E1124 09:13:29.817056 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerName="nova-api-log" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.817116 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerName="nova-api-log" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.817365 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerName="nova-api-log" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.817571 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" containerName="nova-api-api" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.818660 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.821490 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.821687 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.823911 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.829925 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.949285 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-public-tls-certs\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.949338 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.949384 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh9gq\" (UniqueName: \"kubernetes.io/projected/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-kube-api-access-qh9gq\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.949984 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-logs\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.950038 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.950129 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-config-data\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.950386 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-96smc"] Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.952261 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.954358 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.954523 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 09:13:29 crc kubenswrapper[4944]: I1124 09:13:29.966287 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-96smc"] Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.052405 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.052755 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh9gq\" (UniqueName: \"kubernetes.io/projected/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-kube-api-access-qh9gq\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.052926 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-logs\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.053094 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.053257 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-config-data\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.053464 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-public-tls-certs\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.053619 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-logs\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.057542 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.057828 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.061641 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-public-tls-certs\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.071482 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-config-data\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.083079 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh9gq\" (UniqueName: \"kubernetes.io/projected/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-kube-api-access-qh9gq\") pod \"nova-api-0\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.147615 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.154656 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-scripts\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.154873 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d84m9\" (UniqueName: \"kubernetes.io/projected/836f8fae-73e8-423e-8dec-0ee779d09985-kube-api-access-d84m9\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.154985 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-config-data\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.155140 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.267437 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.267766 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-scripts\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.267980 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d84m9\" (UniqueName: \"kubernetes.io/projected/836f8fae-73e8-423e-8dec-0ee779d09985-kube-api-access-d84m9\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.268031 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-config-data\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.272189 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.273681 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-config-data\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.274602 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-scripts\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.288468 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d84m9\" (UniqueName: \"kubernetes.io/projected/836f8fae-73e8-423e-8dec-0ee779d09985-kube-api-access-d84m9\") pod \"nova-cell1-cell-mapping-96smc\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.300752 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23ad5a1c-7ea6-4909-a783-b14b80a4bb7d" path="/var/lib/kubelet/pods/23ad5a1c-7ea6-4909-a783-b14b80a4bb7d/volumes" Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.316821 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:30 crc kubenswrapper[4944]: W1124 09:13:30.626720 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4ba73a4_893e_4b28_a11b_23cdae55f7f1.slice/crio-029d34247b3ca4b2771c2f4bff5c30f3b925acba531f28db1dbf57f832e6c97f WatchSource:0}: Error finding container 029d34247b3ca4b2771c2f4bff5c30f3b925acba531f28db1dbf57f832e6c97f: Status 404 returned error can't find the container with id 029d34247b3ca4b2771c2f4bff5c30f3b925acba531f28db1dbf57f832e6c97f Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.630393 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.729011 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4","Type":"ContainerStarted","Data":"dbee395d02885c9a06d22bd217f37d54477844b3e398fa6029fcb08eebb7be4e"} Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.730275 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4ba73a4-893e-4b28-a11b-23cdae55f7f1","Type":"ContainerStarted","Data":"029d34247b3ca4b2771c2f4bff5c30f3b925acba531f28db1dbf57f832e6c97f"} Nov 24 09:13:30 crc kubenswrapper[4944]: I1124 09:13:30.752498 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-96smc"] Nov 24 09:13:30 crc kubenswrapper[4944]: W1124 09:13:30.758206 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod836f8fae_73e8_423e_8dec_0ee779d09985.slice/crio-e07bbc16daf93ee96c871667a398da80e713613c229bd474bc7176554b992013 WatchSource:0}: Error finding container e07bbc16daf93ee96c871667a398da80e713613c229bd474bc7176554b992013: Status 404 returned error can't find the container with id e07bbc16daf93ee96c871667a398da80e713613c229bd474bc7176554b992013 Nov 24 09:13:31 crc kubenswrapper[4944]: I1124 09:13:31.748470 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96smc" event={"ID":"836f8fae-73e8-423e-8dec-0ee779d09985","Type":"ContainerStarted","Data":"e427978bf154c1166e0f3fa4b8da09e0203655a4a66eff43c601c65fe8459316"} Nov 24 09:13:31 crc kubenswrapper[4944]: I1124 09:13:31.748980 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96smc" event={"ID":"836f8fae-73e8-423e-8dec-0ee779d09985","Type":"ContainerStarted","Data":"e07bbc16daf93ee96c871667a398da80e713613c229bd474bc7176554b992013"} Nov 24 09:13:31 crc kubenswrapper[4944]: I1124 09:13:31.752357 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4ba73a4-893e-4b28-a11b-23cdae55f7f1","Type":"ContainerStarted","Data":"0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908"} Nov 24 09:13:31 crc kubenswrapper[4944]: I1124 09:13:31.752404 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4ba73a4-893e-4b28-a11b-23cdae55f7f1","Type":"ContainerStarted","Data":"84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366"} Nov 24 09:13:31 crc kubenswrapper[4944]: I1124 09:13:31.763158 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-96smc" podStartSLOduration=2.763136932 podStartE2EDuration="2.763136932s" podCreationTimestamp="2025-11-24 09:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:13:31.760524038 +0000 UTC m=+1272.294964510" watchObservedRunningTime="2025-11-24 09:13:31.763136932 +0000 UTC m=+1272.297577394" Nov 24 09:13:31 crc kubenswrapper[4944]: I1124 09:13:31.781550 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.781528954 podStartE2EDuration="2.781528954s" podCreationTimestamp="2025-11-24 09:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:13:31.777799165 +0000 UTC m=+1272.312239647" watchObservedRunningTime="2025-11-24 09:13:31.781528954 +0000 UTC m=+1272.315969416" Nov 24 09:13:32 crc kubenswrapper[4944]: I1124 09:13:32.268245 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:13:32 crc kubenswrapper[4944]: I1124 09:13:32.363846 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-8wv9x"] Nov 24 09:13:32 crc kubenswrapper[4944]: I1124 09:13:32.366506 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" podUID="e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" containerName="dnsmasq-dns" containerID="cri-o://577374ee469d4a9c77f3adeb7bc1d311545ffbc6d279cd8e1f90d2a17c173257" gracePeriod=10 Nov 24 09:13:32 crc kubenswrapper[4944]: I1124 09:13:32.764380 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4","Type":"ContainerStarted","Data":"a84b239223662f1225b37375264550577380c4e7d01f7c269035f7022dd18d82"} Nov 24 09:13:32 crc kubenswrapper[4944]: I1124 09:13:32.765535 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 09:13:32 crc kubenswrapper[4944]: I1124 09:13:32.779482 4944 generic.go:334] "Generic (PLEG): container finished" podID="e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" containerID="577374ee469d4a9c77f3adeb7bc1d311545ffbc6d279cd8e1f90d2a17c173257" exitCode=0 Nov 24 09:13:32 crc kubenswrapper[4944]: I1124 09:13:32.780419 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" event={"ID":"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0","Type":"ContainerDied","Data":"577374ee469d4a9c77f3adeb7bc1d311545ffbc6d279cd8e1f90d2a17c173257"} Nov 24 09:13:32 crc kubenswrapper[4944]: I1124 09:13:32.803789 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.71248668 podStartE2EDuration="6.803767865s" podCreationTimestamp="2025-11-24 09:13:26 +0000 UTC" firstStartedPulling="2025-11-24 09:13:27.477713159 +0000 UTC m=+1268.012153621" lastFinishedPulling="2025-11-24 09:13:31.568994334 +0000 UTC m=+1272.103434806" observedRunningTime="2025-11-24 09:13:32.797231975 +0000 UTC m=+1273.331672447" watchObservedRunningTime="2025-11-24 09:13:32.803767865 +0000 UTC m=+1273.338208327" Nov 24 09:13:32 crc kubenswrapper[4944]: I1124 09:13:32.942881 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.147683 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-nb\") pod \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.148430 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-sb\") pod \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.148467 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-svc\") pod \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.148484 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnmnl\" (UniqueName: \"kubernetes.io/projected/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-kube-api-access-xnmnl\") pod \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.148503 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-swift-storage-0\") pod \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.148548 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-config\") pod \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\" (UID: \"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0\") " Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.159520 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-kube-api-access-xnmnl" (OuterVolumeSpecName: "kube-api-access-xnmnl") pod "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" (UID: "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0"). InnerVolumeSpecName "kube-api-access-xnmnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.203257 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" (UID: "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.210404 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" (UID: "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.212547 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-config" (OuterVolumeSpecName: "config") pod "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" (UID: "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.214779 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" (UID: "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.226586 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" (UID: "e8b63713-ae4a-4cb3-98f7-0a51c7924ff0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.258444 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.258498 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.258514 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.258529 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnmnl\" (UniqueName: \"kubernetes.io/projected/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-kube-api-access-xnmnl\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.258545 4944 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.258556 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.790866 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" event={"ID":"e8b63713-ae4a-4cb3-98f7-0a51c7924ff0","Type":"ContainerDied","Data":"a776c62271e6aa1c29d87621bd1053914dfdcfc4171f7ff0770ce52af58a1362"} Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.790892 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-8wv9x" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.790935 4944 scope.go:117] "RemoveContainer" containerID="577374ee469d4a9c77f3adeb7bc1d311545ffbc6d279cd8e1f90d2a17c173257" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.827180 4944 scope.go:117] "RemoveContainer" containerID="f0b36825e86a4f2197ef7bde88dbfd675a16b202341b45e48fc5bf1000094669" Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.831110 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-8wv9x"] Nov 24 09:13:33 crc kubenswrapper[4944]: I1124 09:13:33.839378 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-8wv9x"] Nov 24 09:13:34 crc kubenswrapper[4944]: I1124 09:13:34.302546 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" path="/var/lib/kubelet/pods/e8b63713-ae4a-4cb3-98f7-0a51c7924ff0/volumes" Nov 24 09:13:36 crc kubenswrapper[4944]: I1124 09:13:36.830089 4944 generic.go:334] "Generic (PLEG): container finished" podID="836f8fae-73e8-423e-8dec-0ee779d09985" containerID="e427978bf154c1166e0f3fa4b8da09e0203655a4a66eff43c601c65fe8459316" exitCode=0 Nov 24 09:13:36 crc kubenswrapper[4944]: I1124 09:13:36.830174 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96smc" event={"ID":"836f8fae-73e8-423e-8dec-0ee779d09985","Type":"ContainerDied","Data":"e427978bf154c1166e0f3fa4b8da09e0203655a4a66eff43c601c65fe8459316"} Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.185112 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.375260 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-combined-ca-bundle\") pod \"836f8fae-73e8-423e-8dec-0ee779d09985\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.375429 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-scripts\") pod \"836f8fae-73e8-423e-8dec-0ee779d09985\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.375458 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-config-data\") pod \"836f8fae-73e8-423e-8dec-0ee779d09985\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.375746 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d84m9\" (UniqueName: \"kubernetes.io/projected/836f8fae-73e8-423e-8dec-0ee779d09985-kube-api-access-d84m9\") pod \"836f8fae-73e8-423e-8dec-0ee779d09985\" (UID: \"836f8fae-73e8-423e-8dec-0ee779d09985\") " Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.381111 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-scripts" (OuterVolumeSpecName: "scripts") pod "836f8fae-73e8-423e-8dec-0ee779d09985" (UID: "836f8fae-73e8-423e-8dec-0ee779d09985"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.381387 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/836f8fae-73e8-423e-8dec-0ee779d09985-kube-api-access-d84m9" (OuterVolumeSpecName: "kube-api-access-d84m9") pod "836f8fae-73e8-423e-8dec-0ee779d09985" (UID: "836f8fae-73e8-423e-8dec-0ee779d09985"). InnerVolumeSpecName "kube-api-access-d84m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.402146 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-config-data" (OuterVolumeSpecName: "config-data") pod "836f8fae-73e8-423e-8dec-0ee779d09985" (UID: "836f8fae-73e8-423e-8dec-0ee779d09985"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.402639 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "836f8fae-73e8-423e-8dec-0ee779d09985" (UID: "836f8fae-73e8-423e-8dec-0ee779d09985"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.479262 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.479297 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.479309 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/836f8fae-73e8-423e-8dec-0ee779d09985-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.479318 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d84m9\" (UniqueName: \"kubernetes.io/projected/836f8fae-73e8-423e-8dec-0ee779d09985-kube-api-access-d84m9\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.848169 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96smc" event={"ID":"836f8fae-73e8-423e-8dec-0ee779d09985","Type":"ContainerDied","Data":"e07bbc16daf93ee96c871667a398da80e713613c229bd474bc7176554b992013"} Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.848207 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e07bbc16daf93ee96c871667a398da80e713613c229bd474bc7176554b992013" Nov 24 09:13:38 crc kubenswrapper[4944]: I1124 09:13:38.848243 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96smc" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.026145 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.026400 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c4ba73a4-893e-4b28-a11b-23cdae55f7f1" containerName="nova-api-log" containerID="cri-o://84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366" gracePeriod=30 Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.026714 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c4ba73a4-893e-4b28-a11b-23cdae55f7f1" containerName="nova-api-api" containerID="cri-o://0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908" gracePeriod=30 Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.051152 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.051766 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c45f64f4-e570-4379-9037-d8cf5712ee20" containerName="nova-scheduler-scheduler" containerID="cri-o://6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d" gracePeriod=30 Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.061077 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.061310 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-log" containerID="cri-o://d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0" gracePeriod=30 Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.061708 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-metadata" containerID="cri-o://09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18" gracePeriod=30 Nov 24 09:13:39 crc kubenswrapper[4944]: E1124 09:13:39.231167 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbae34b6c_03f3_4c99_b7c5_400541549ed3.slice/crio-d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4ba73a4_893e_4b28_a11b_23cdae55f7f1.slice/crio-84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbae34b6c_03f3_4c99_b7c5_400541549ed3.slice/crio-conmon-d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0.scope\": RecentStats: unable to find data in memory cache]" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.690263 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.804744 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-combined-ca-bundle\") pod \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.804883 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh9gq\" (UniqueName: \"kubernetes.io/projected/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-kube-api-access-qh9gq\") pod \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.804951 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-internal-tls-certs\") pod \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.804995 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-config-data\") pod \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.805097 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-logs\") pod \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.805163 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-public-tls-certs\") pod \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\" (UID: \"c4ba73a4-893e-4b28-a11b-23cdae55f7f1\") " Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.807064 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-logs" (OuterVolumeSpecName: "logs") pod "c4ba73a4-893e-4b28-a11b-23cdae55f7f1" (UID: "c4ba73a4-893e-4b28-a11b-23cdae55f7f1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.810794 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-kube-api-access-qh9gq" (OuterVolumeSpecName: "kube-api-access-qh9gq") pod "c4ba73a4-893e-4b28-a11b-23cdae55f7f1" (UID: "c4ba73a4-893e-4b28-a11b-23cdae55f7f1"). InnerVolumeSpecName "kube-api-access-qh9gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.836722 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-config-data" (OuterVolumeSpecName: "config-data") pod "c4ba73a4-893e-4b28-a11b-23cdae55f7f1" (UID: "c4ba73a4-893e-4b28-a11b-23cdae55f7f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.839122 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4ba73a4-893e-4b28-a11b-23cdae55f7f1" (UID: "c4ba73a4-893e-4b28-a11b-23cdae55f7f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.861391 4944 generic.go:334] "Generic (PLEG): container finished" podID="c4ba73a4-893e-4b28-a11b-23cdae55f7f1" containerID="0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908" exitCode=0 Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.861442 4944 generic.go:334] "Generic (PLEG): container finished" podID="c4ba73a4-893e-4b28-a11b-23cdae55f7f1" containerID="84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366" exitCode=143 Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.861506 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4ba73a4-893e-4b28-a11b-23cdae55f7f1","Type":"ContainerDied","Data":"0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908"} Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.861572 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4ba73a4-893e-4b28-a11b-23cdae55f7f1","Type":"ContainerDied","Data":"84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366"} Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.861589 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c4ba73a4-893e-4b28-a11b-23cdae55f7f1","Type":"ContainerDied","Data":"029d34247b3ca4b2771c2f4bff5c30f3b925acba531f28db1dbf57f832e6c97f"} Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.861615 4944 scope.go:117] "RemoveContainer" containerID="0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.861512 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.864575 4944 generic.go:334] "Generic (PLEG): container finished" podID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerID="d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0" exitCode=143 Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.864610 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bae34b6c-03f3-4c99-b7c5-400541549ed3","Type":"ContainerDied","Data":"d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0"} Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.871183 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c4ba73a4-893e-4b28-a11b-23cdae55f7f1" (UID: "c4ba73a4-893e-4b28-a11b-23cdae55f7f1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.871825 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c4ba73a4-893e-4b28-a11b-23cdae55f7f1" (UID: "c4ba73a4-893e-4b28-a11b-23cdae55f7f1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.888510 4944 scope.go:117] "RemoveContainer" containerID="84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.911341 4944 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.911397 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.911413 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.911427 4944 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.911439 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.911449 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh9gq\" (UniqueName: \"kubernetes.io/projected/c4ba73a4-893e-4b28-a11b-23cdae55f7f1-kube-api-access-qh9gq\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.913291 4944 scope.go:117] "RemoveContainer" containerID="0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908" Nov 24 09:13:39 crc kubenswrapper[4944]: E1124 09:13:39.913839 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908\": container with ID starting with 0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908 not found: ID does not exist" containerID="0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.913878 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908"} err="failed to get container status \"0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908\": rpc error: code = NotFound desc = could not find container \"0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908\": container with ID starting with 0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908 not found: ID does not exist" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.913907 4944 scope.go:117] "RemoveContainer" containerID="84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366" Nov 24 09:13:39 crc kubenswrapper[4944]: E1124 09:13:39.914737 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366\": container with ID starting with 84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366 not found: ID does not exist" containerID="84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.914767 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366"} err="failed to get container status \"84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366\": rpc error: code = NotFound desc = could not find container \"84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366\": container with ID starting with 84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366 not found: ID does not exist" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.914781 4944 scope.go:117] "RemoveContainer" containerID="0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.915242 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908"} err="failed to get container status \"0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908\": rpc error: code = NotFound desc = could not find container \"0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908\": container with ID starting with 0a9a27d10e669039fb902e8c4041d02ee2cd11d918b23e511cc150a5b7273908 not found: ID does not exist" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.915287 4944 scope.go:117] "RemoveContainer" containerID="84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366" Nov 24 09:13:39 crc kubenswrapper[4944]: I1124 09:13:39.915582 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366"} err="failed to get container status \"84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366\": rpc error: code = NotFound desc = could not find container \"84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366\": container with ID starting with 84fb455ef8827833f6610ceaf403b79f69322f8362c622cf2f5063f21527d366 not found: ID does not exist" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.215284 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.226417 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.243066 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:40 crc kubenswrapper[4944]: E1124 09:13:40.243866 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ba73a4-893e-4b28-a11b-23cdae55f7f1" containerName="nova-api-api" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.243881 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ba73a4-893e-4b28-a11b-23cdae55f7f1" containerName="nova-api-api" Nov 24 09:13:40 crc kubenswrapper[4944]: E1124 09:13:40.243892 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" containerName="init" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.243899 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" containerName="init" Nov 24 09:13:40 crc kubenswrapper[4944]: E1124 09:13:40.243914 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="836f8fae-73e8-423e-8dec-0ee779d09985" containerName="nova-manage" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.243920 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="836f8fae-73e8-423e-8dec-0ee779d09985" containerName="nova-manage" Nov 24 09:13:40 crc kubenswrapper[4944]: E1124 09:13:40.243933 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" containerName="dnsmasq-dns" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.243938 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" containerName="dnsmasq-dns" Nov 24 09:13:40 crc kubenswrapper[4944]: E1124 09:13:40.244003 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ba73a4-893e-4b28-a11b-23cdae55f7f1" containerName="nova-api-log" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.244009 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ba73a4-893e-4b28-a11b-23cdae55f7f1" containerName="nova-api-log" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.244204 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ba73a4-893e-4b28-a11b-23cdae55f7f1" containerName="nova-api-api" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.244234 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ba73a4-893e-4b28-a11b-23cdae55f7f1" containerName="nova-api-log" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.244247 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8b63713-ae4a-4cb3-98f7-0a51c7924ff0" containerName="dnsmasq-dns" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.244262 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="836f8fae-73e8-423e-8dec-0ee779d09985" containerName="nova-manage" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.245287 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.247292 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.247699 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.248929 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.255510 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.297389 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4ba73a4-893e-4b28-a11b-23cdae55f7f1" path="/var/lib/kubelet/pods/c4ba73a4-893e-4b28-a11b-23cdae55f7f1/volumes" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.419843 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-config-data\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.419905 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-public-tls-certs\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.420074 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9kp5\" (UniqueName: \"kubernetes.io/projected/ce23facc-1a0d-4c53-b696-b37a7525eba7-kube-api-access-z9kp5\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.420135 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce23facc-1a0d-4c53-b696-b37a7525eba7-logs\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.420177 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.420196 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.521822 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9kp5\" (UniqueName: \"kubernetes.io/projected/ce23facc-1a0d-4c53-b696-b37a7525eba7-kube-api-access-z9kp5\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.522240 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce23facc-1a0d-4c53-b696-b37a7525eba7-logs\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.522283 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.522298 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.522330 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-config-data\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.522354 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-public-tls-certs\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.523336 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce23facc-1a0d-4c53-b696-b37a7525eba7-logs\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.527314 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.527709 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-config-data\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.528723 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-public-tls-certs\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.533709 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.538784 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9kp5\" (UniqueName: \"kubernetes.io/projected/ce23facc-1a0d-4c53-b696-b37a7525eba7-kube-api-access-z9kp5\") pod \"nova-api-0\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.561640 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.824534 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.897284 4944 generic.go:334] "Generic (PLEG): container finished" podID="c45f64f4-e570-4379-9037-d8cf5712ee20" containerID="6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d" exitCode=0 Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.897378 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.897369 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c45f64f4-e570-4379-9037-d8cf5712ee20","Type":"ContainerDied","Data":"6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d"} Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.897525 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c45f64f4-e570-4379-9037-d8cf5712ee20","Type":"ContainerDied","Data":"e068be7bb5221a001225249f3c776ec49f1e05e9907f6a105757adb096947caf"} Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.897545 4944 scope.go:117] "RemoveContainer" containerID="6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.923273 4944 scope.go:117] "RemoveContainer" containerID="6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d" Nov 24 09:13:40 crc kubenswrapper[4944]: E1124 09:13:40.923799 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d\": container with ID starting with 6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d not found: ID does not exist" containerID="6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.923865 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d"} err="failed to get container status \"6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d\": rpc error: code = NotFound desc = could not find container \"6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d\": container with ID starting with 6d328385f727cecb7c5ffbf0f756fde1879b69c2c885603ea38bcb3a92a8030d not found: ID does not exist" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.943734 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-config-data\") pod \"c45f64f4-e570-4379-9037-d8cf5712ee20\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.943820 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vxzw\" (UniqueName: \"kubernetes.io/projected/c45f64f4-e570-4379-9037-d8cf5712ee20-kube-api-access-9vxzw\") pod \"c45f64f4-e570-4379-9037-d8cf5712ee20\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.943855 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-combined-ca-bundle\") pod \"c45f64f4-e570-4379-9037-d8cf5712ee20\" (UID: \"c45f64f4-e570-4379-9037-d8cf5712ee20\") " Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.951137 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c45f64f4-e570-4379-9037-d8cf5712ee20-kube-api-access-9vxzw" (OuterVolumeSpecName: "kube-api-access-9vxzw") pod "c45f64f4-e570-4379-9037-d8cf5712ee20" (UID: "c45f64f4-e570-4379-9037-d8cf5712ee20"). InnerVolumeSpecName "kube-api-access-9vxzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.981657 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c45f64f4-e570-4379-9037-d8cf5712ee20" (UID: "c45f64f4-e570-4379-9037-d8cf5712ee20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:40 crc kubenswrapper[4944]: I1124 09:13:40.987115 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-config-data" (OuterVolumeSpecName: "config-data") pod "c45f64f4-e570-4379-9037-d8cf5712ee20" (UID: "c45f64f4-e570-4379-9037-d8cf5712ee20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.046974 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.047017 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vxzw\" (UniqueName: \"kubernetes.io/projected/c45f64f4-e570-4379-9037-d8cf5712ee20-kube-api-access-9vxzw\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.047028 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c45f64f4-e570-4379-9037-d8cf5712ee20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.062201 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:13:41 crc kubenswrapper[4944]: W1124 09:13:41.062680 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce23facc_1a0d_4c53_b696_b37a7525eba7.slice/crio-f47f97b42ee843b2530865726437b567dd13721da1dc0af7e10f87abd25997a5 WatchSource:0}: Error finding container f47f97b42ee843b2530865726437b567dd13721da1dc0af7e10f87abd25997a5: Status 404 returned error can't find the container with id f47f97b42ee843b2530865726437b567dd13721da1dc0af7e10f87abd25997a5 Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.230423 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.240745 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.252578 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:13:41 crc kubenswrapper[4944]: E1124 09:13:41.253057 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c45f64f4-e570-4379-9037-d8cf5712ee20" containerName="nova-scheduler-scheduler" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.253071 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c45f64f4-e570-4379-9037-d8cf5712ee20" containerName="nova-scheduler-scheduler" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.253242 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="c45f64f4-e570-4379-9037-d8cf5712ee20" containerName="nova-scheduler-scheduler" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.253850 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.259396 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.261079 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.357475 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.357642 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvbj6\" (UniqueName: \"kubernetes.io/projected/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-kube-api-access-hvbj6\") pod \"nova-scheduler-0\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.357674 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-config-data\") pod \"nova-scheduler-0\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.460674 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvbj6\" (UniqueName: \"kubernetes.io/projected/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-kube-api-access-hvbj6\") pod \"nova-scheduler-0\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.460724 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-config-data\") pod \"nova-scheduler-0\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.460814 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.465159 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-config-data\") pod \"nova-scheduler-0\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.465901 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.487655 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvbj6\" (UniqueName: \"kubernetes.io/projected/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-kube-api-access-hvbj6\") pod \"nova-scheduler-0\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " pod="openstack/nova-scheduler-0" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.576931 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.909649 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce23facc-1a0d-4c53-b696-b37a7525eba7","Type":"ContainerStarted","Data":"1156a54f88c1c2dd0577bc2b082f7256b862e090c82e8172fc3df6b5bf9e5bdf"} Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.909979 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce23facc-1a0d-4c53-b696-b37a7525eba7","Type":"ContainerStarted","Data":"ceef1855e41bb1981a2f6e513be92fa9f3b4e90878c9354637a799ca0b3bbbf4"} Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.909990 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce23facc-1a0d-4c53-b696-b37a7525eba7","Type":"ContainerStarted","Data":"f47f97b42ee843b2530865726437b567dd13721da1dc0af7e10f87abd25997a5"} Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.929985 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.929966032 podStartE2EDuration="1.929966032s" podCreationTimestamp="2025-11-24 09:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:13:41.925729906 +0000 UTC m=+1282.460170378" watchObservedRunningTime="2025-11-24 09:13:41.929966032 +0000 UTC m=+1282.464406494" Nov 24 09:13:41 crc kubenswrapper[4944]: I1124 09:13:41.999221 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.218639 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:48624->10.217.0.189:8775: read: connection reset by peer" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.218731 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:48614->10.217.0.189:8775: read: connection reset by peer" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.293641 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c45f64f4-e570-4379-9037-d8cf5712ee20" path="/var/lib/kubelet/pods/c45f64f4-e570-4379-9037-d8cf5712ee20/volumes" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.668065 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.784272 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-nova-metadata-tls-certs\") pod \"bae34b6c-03f3-4c99-b7c5-400541549ed3\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.785174 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-config-data\") pod \"bae34b6c-03f3-4c99-b7c5-400541549ed3\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.785238 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-combined-ca-bundle\") pod \"bae34b6c-03f3-4c99-b7c5-400541549ed3\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.785280 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvpmp\" (UniqueName: \"kubernetes.io/projected/bae34b6c-03f3-4c99-b7c5-400541549ed3-kube-api-access-vvpmp\") pod \"bae34b6c-03f3-4c99-b7c5-400541549ed3\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.785397 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bae34b6c-03f3-4c99-b7c5-400541549ed3-logs\") pod \"bae34b6c-03f3-4c99-b7c5-400541549ed3\" (UID: \"bae34b6c-03f3-4c99-b7c5-400541549ed3\") " Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.786841 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bae34b6c-03f3-4c99-b7c5-400541549ed3-logs" (OuterVolumeSpecName: "logs") pod "bae34b6c-03f3-4c99-b7c5-400541549ed3" (UID: "bae34b6c-03f3-4c99-b7c5-400541549ed3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.793655 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae34b6c-03f3-4c99-b7c5-400541549ed3-kube-api-access-vvpmp" (OuterVolumeSpecName: "kube-api-access-vvpmp") pod "bae34b6c-03f3-4c99-b7c5-400541549ed3" (UID: "bae34b6c-03f3-4c99-b7c5-400541549ed3"). InnerVolumeSpecName "kube-api-access-vvpmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.820225 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bae34b6c-03f3-4c99-b7c5-400541549ed3" (UID: "bae34b6c-03f3-4c99-b7c5-400541549ed3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.830321 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-config-data" (OuterVolumeSpecName: "config-data") pod "bae34b6c-03f3-4c99-b7c5-400541549ed3" (UID: "bae34b6c-03f3-4c99-b7c5-400541549ed3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.867486 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "bae34b6c-03f3-4c99-b7c5-400541549ed3" (UID: "bae34b6c-03f3-4c99-b7c5-400541549ed3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.889004 4944 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.889274 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.889286 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae34b6c-03f3-4c99-b7c5-400541549ed3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.889295 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvpmp\" (UniqueName: \"kubernetes.io/projected/bae34b6c-03f3-4c99-b7c5-400541549ed3-kube-api-access-vvpmp\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.889303 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bae34b6c-03f3-4c99-b7c5-400541549ed3-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.919748 4944 generic.go:334] "Generic (PLEG): container finished" podID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerID="09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18" exitCode=0 Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.919824 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bae34b6c-03f3-4c99-b7c5-400541549ed3","Type":"ContainerDied","Data":"09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18"} Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.919856 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bae34b6c-03f3-4c99-b7c5-400541549ed3","Type":"ContainerDied","Data":"04cde4650ac0dd5c85edff384cee6a87a8526b6865fb44bd27474cd0d27b0b7b"} Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.919875 4944 scope.go:117] "RemoveContainer" containerID="09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.919993 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.933671 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"90fb3cc2-0725-48dc-9ad5-5c8da27930f8","Type":"ContainerStarted","Data":"58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249"} Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.933707 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"90fb3cc2-0725-48dc-9ad5-5c8da27930f8","Type":"ContainerStarted","Data":"28935e94026f788d13d8f73bdafe9c288381d3873af32802f62205a576bebfee"} Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.954807 4944 scope.go:117] "RemoveContainer" containerID="d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.957265 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.957245074 podStartE2EDuration="1.957245074s" podCreationTimestamp="2025-11-24 09:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:13:42.95681056 +0000 UTC m=+1283.491251032" watchObservedRunningTime="2025-11-24 09:13:42.957245074 +0000 UTC m=+1283.491685536" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.979404 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.986259 4944 scope.go:117] "RemoveContainer" containerID="09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18" Nov 24 09:13:42 crc kubenswrapper[4944]: E1124 09:13:42.986746 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18\": container with ID starting with 09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18 not found: ID does not exist" containerID="09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.986783 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18"} err="failed to get container status \"09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18\": rpc error: code = NotFound desc = could not find container \"09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18\": container with ID starting with 09be08930db5fdbfee09af39f2a612ea58245068fa4e741c478333bbfa39bb18 not found: ID does not exist" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.986808 4944 scope.go:117] "RemoveContainer" containerID="d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0" Nov 24 09:13:42 crc kubenswrapper[4944]: E1124 09:13:42.987270 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0\": container with ID starting with d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0 not found: ID does not exist" containerID="d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0" Nov 24 09:13:42 crc kubenswrapper[4944]: I1124 09:13:42.987323 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0"} err="failed to get container status \"d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0\": rpc error: code = NotFound desc = could not find container \"d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0\": container with ID starting with d507ff856ad9723fd04e7eba7798d421b5ada4534258865824cf75ab510e76f0 not found: ID does not exist" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.005008 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.013274 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:13:43 crc kubenswrapper[4944]: E1124 09:13:43.013753 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-metadata" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.013772 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-metadata" Nov 24 09:13:43 crc kubenswrapper[4944]: E1124 09:13:43.013823 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-log" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.015076 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-log" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.015361 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-metadata" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.015394 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" containerName="nova-metadata-log" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.016964 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.019971 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.020240 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.023134 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.194468 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.194536 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-config-data\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.194583 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.194616 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc12ffc5-e026-42a6-8ddc-43961c535787-logs\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.194773 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqqx7\" (UniqueName: \"kubernetes.io/projected/fc12ffc5-e026-42a6-8ddc-43961c535787-kube-api-access-zqqx7\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.299151 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqqx7\" (UniqueName: \"kubernetes.io/projected/fc12ffc5-e026-42a6-8ddc-43961c535787-kube-api-access-zqqx7\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.299224 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.299254 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-config-data\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.299337 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.299372 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc12ffc5-e026-42a6-8ddc-43961c535787-logs\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.299859 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc12ffc5-e026-42a6-8ddc-43961c535787-logs\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.303098 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.308675 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.313451 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-config-data\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.327942 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqqx7\" (UniqueName: \"kubernetes.io/projected/fc12ffc5-e026-42a6-8ddc-43961c535787-kube-api-access-zqqx7\") pod \"nova-metadata-0\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.335549 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.769908 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.944514 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc12ffc5-e026-42a6-8ddc-43961c535787","Type":"ContainerStarted","Data":"3dfe8ae924d36e9fb2dbf24bc72c97e4e845aaa92a56c64d0448011249204cfb"} Nov 24 09:13:43 crc kubenswrapper[4944]: I1124 09:13:43.944567 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc12ffc5-e026-42a6-8ddc-43961c535787","Type":"ContainerStarted","Data":"6244c6b898aad9686ec3893884879e2cc73ae2fbb8dc7c6685421ec110803fa9"} Nov 24 09:13:44 crc kubenswrapper[4944]: I1124 09:13:44.286828 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bae34b6c-03f3-4c99-b7c5-400541549ed3" path="/var/lib/kubelet/pods/bae34b6c-03f3-4c99-b7c5-400541549ed3/volumes" Nov 24 09:13:44 crc kubenswrapper[4944]: I1124 09:13:44.960909 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc12ffc5-e026-42a6-8ddc-43961c535787","Type":"ContainerStarted","Data":"f4fda6a5047a5412abc68148082d7806e36494cd3997be68245586f663ec9996"} Nov 24 09:13:44 crc kubenswrapper[4944]: I1124 09:13:44.991724 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.991696402 podStartE2EDuration="2.991696402s" podCreationTimestamp="2025-11-24 09:13:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:13:44.978858977 +0000 UTC m=+1285.513299459" watchObservedRunningTime="2025-11-24 09:13:44.991696402 +0000 UTC m=+1285.526136884" Nov 24 09:13:46 crc kubenswrapper[4944]: I1124 09:13:46.577348 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 09:13:48 crc kubenswrapper[4944]: I1124 09:13:48.335985 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 09:13:48 crc kubenswrapper[4944]: I1124 09:13:48.336335 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 09:13:50 crc kubenswrapper[4944]: I1124 09:13:50.562582 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 09:13:50 crc kubenswrapper[4944]: I1124 09:13:50.562917 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 09:13:51 crc kubenswrapper[4944]: I1124 09:13:51.577814 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 09:13:51 crc kubenswrapper[4944]: I1124 09:13:51.578496 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 09:13:51 crc kubenswrapper[4944]: I1124 09:13:51.579249 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 09:13:51 crc kubenswrapper[4944]: I1124 09:13:51.605638 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 09:13:52 crc kubenswrapper[4944]: I1124 09:13:52.047112 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 09:13:53 crc kubenswrapper[4944]: I1124 09:13:53.335723 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 09:13:53 crc kubenswrapper[4944]: I1124 09:13:53.336089 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 09:13:53 crc kubenswrapper[4944]: I1124 09:13:53.548415 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:13:53 crc kubenswrapper[4944]: I1124 09:13:53.549065 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:13:54 crc kubenswrapper[4944]: I1124 09:13:54.350198 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 09:13:54 crc kubenswrapper[4944]: I1124 09:13:54.350198 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 09:13:57 crc kubenswrapper[4944]: I1124 09:13:57.051701 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 09:14:00 crc kubenswrapper[4944]: I1124 09:14:00.568673 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 09:14:00 crc kubenswrapper[4944]: I1124 09:14:00.570207 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 09:14:00 crc kubenswrapper[4944]: I1124 09:14:00.570871 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 09:14:00 crc kubenswrapper[4944]: I1124 09:14:00.579743 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 09:14:01 crc kubenswrapper[4944]: I1124 09:14:01.096446 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 09:14:01 crc kubenswrapper[4944]: I1124 09:14:01.104315 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 09:14:03 crc kubenswrapper[4944]: I1124 09:14:03.340503 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 09:14:03 crc kubenswrapper[4944]: I1124 09:14:03.340840 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 09:14:03 crc kubenswrapper[4944]: I1124 09:14:03.346368 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 09:14:03 crc kubenswrapper[4944]: I1124 09:14:03.346761 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.534917 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.567290 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.567517 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="9a8dab89-ec08-4497-a890-bcb70c8ea430" containerName="openstackclient" containerID="cri-o://194effb1be878a12474f5aad321c9b034378e690429268d01ba3188c16709667" gracePeriod=2 Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.585291 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 09:14:20 crc kubenswrapper[4944]: E1124 09:14:20.631532 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 09:14:20 crc kubenswrapper[4944]: E1124 09:14:20.631908 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data podName:67e92d3f-3532-467f-bf38-c9c3107d4fcb nodeName:}" failed. No retries permitted until 2025-11-24 09:14:21.13188644 +0000 UTC m=+1321.666326902 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data") pod "rabbitmq-server-0" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb") : configmap "rabbitmq-config-data" not found Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.708224 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicand270-account-delete-h2tcc"] Nov 24 09:14:20 crc kubenswrapper[4944]: E1124 09:14:20.708805 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a8dab89-ec08-4497-a890-bcb70c8ea430" containerName="openstackclient" Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.708825 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a8dab89-ec08-4497-a890-bcb70c8ea430" containerName="openstackclient" Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.709207 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a8dab89-ec08-4497-a890-bcb70c8ea430" containerName="openstackclient" Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.710125 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicand270-account-delete-h2tcc" Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.746315 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.746807 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerName="ovn-northd" containerID="cri-o://5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c" gracePeriod=30 Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.746875 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerName="openstack-network-exporter" containerID="cri-o://81af0f39619d4f0d4399fe088165e6c6fb343c03871f10888e1d787a30b8aa80" gracePeriod=30 Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.767109 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicand270-account-delete-h2tcc"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.779393 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-npmfc"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.792288 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-m2b6x"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.792569 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-m2b6x" podUID="c521e9a3-0d64-4658-8a00-690d20e619d0" containerName="openstack-network-exporter" containerID="cri-o://b41e32588963f3e10325bdeeaf1f1bb9e18dd5fa9618690b49d86fb5f1d1c724" gracePeriod=30 Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.840075 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-7gtjj"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.842549 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq5x6\" (UniqueName: \"kubernetes.io/projected/55a9098c-7257-46f7-888c-c4bb6f9421a5-kube-api-access-xq5x6\") pod \"barbicand270-account-delete-h2tcc\" (UID: \"55a9098c-7257-46f7-888c-c4bb6f9421a5\") " pod="openstack/barbicand270-account-delete-h2tcc" Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.842739 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a9098c-7257-46f7-888c-c4bb6f9421a5-operator-scripts\") pod \"barbicand270-account-delete-h2tcc\" (UID: \"55a9098c-7257-46f7-888c-c4bb6f9421a5\") " pod="openstack/barbicand270-account-delete-h2tcc" Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.871136 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-hpjlj"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.885767 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-nst6k"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.907537 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-hpjlj"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.948012 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-nst6k"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.949253 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq5x6\" (UniqueName: \"kubernetes.io/projected/55a9098c-7257-46f7-888c-c4bb6f9421a5-kube-api-access-xq5x6\") pod \"barbicand270-account-delete-h2tcc\" (UID: \"55a9098c-7257-46f7-888c-c4bb6f9421a5\") " pod="openstack/barbicand270-account-delete-h2tcc" Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.949422 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a9098c-7257-46f7-888c-c4bb6f9421a5-operator-scripts\") pod \"barbicand270-account-delete-h2tcc\" (UID: \"55a9098c-7257-46f7-888c-c4bb6f9421a5\") " pod="openstack/barbicand270-account-delete-h2tcc" Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.950077 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a9098c-7257-46f7-888c-c4bb6f9421a5-operator-scripts\") pod \"barbicand270-account-delete-h2tcc\" (UID: \"55a9098c-7257-46f7-888c-c4bb6f9421a5\") " pod="openstack/barbicand270-account-delete-h2tcc" Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.970422 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder575d-account-delete-8dqqk"] Nov 24 09:14:20 crc kubenswrapper[4944]: I1124 09:14:20.971880 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder575d-account-delete-8dqqk" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.008411 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rsh4b"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.008683 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" podUID="5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" containerName="dnsmasq-dns" containerID="cri-o://823389b569a8fd3fda96971a913dc73c72a4f283c5b7ab6145f8fb0dd9b6cb9e" gracePeriod=10 Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.029172 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq5x6\" (UniqueName: \"kubernetes.io/projected/55a9098c-7257-46f7-888c-c4bb6f9421a5-kube-api-access-xq5x6\") pod \"barbicand270-account-delete-h2tcc\" (UID: \"55a9098c-7257-46f7-888c-c4bb6f9421a5\") " pod="openstack/barbicand270-account-delete-h2tcc" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.039358 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.043833 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicand270-account-delete-h2tcc" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.053257 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-operator-scripts\") pod \"cinder575d-account-delete-8dqqk\" (UID: \"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2\") " pod="openstack/cinder575d-account-delete-8dqqk" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.053374 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z75h\" (UniqueName: \"kubernetes.io/projected/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-kube-api-access-4z75h\") pod \"cinder575d-account-delete-8dqqk\" (UID: \"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2\") " pod="openstack/cinder575d-account-delete-8dqqk" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.064695 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder575d-account-delete-8dqqk"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.130334 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance509f-account-delete-fx5bb"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.131871 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance509f-account-delete-fx5bb" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.150131 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance509f-account-delete-fx5bb"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.155418 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-operator-scripts\") pod \"cinder575d-account-delete-8dqqk\" (UID: \"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2\") " pod="openstack/cinder575d-account-delete-8dqqk" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.155713 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-operator-scripts\") pod \"glance509f-account-delete-fx5bb\" (UID: \"1db23d75-43f2-4bae-9e09-399d0ffb7ed5\") " pod="openstack/glance509f-account-delete-fx5bb" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.155939 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z75h\" (UniqueName: \"kubernetes.io/projected/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-kube-api-access-4z75h\") pod \"cinder575d-account-delete-8dqqk\" (UID: \"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2\") " pod="openstack/cinder575d-account-delete-8dqqk" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.155988 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qs7r\" (UniqueName: \"kubernetes.io/projected/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-kube-api-access-5qs7r\") pod \"glance509f-account-delete-fx5bb\" (UID: \"1db23d75-43f2-4bae-9e09-399d0ffb7ed5\") " pod="openstack/glance509f-account-delete-fx5bb" Nov 24 09:14:21 crc kubenswrapper[4944]: E1124 09:14:21.156694 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 09:14:21 crc kubenswrapper[4944]: E1124 09:14:21.156745 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data podName:67e92d3f-3532-467f-bf38-c9c3107d4fcb nodeName:}" failed. No retries permitted until 2025-11-24 09:14:22.156730478 +0000 UTC m=+1322.691170940 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data") pod "rabbitmq-server-0" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb") : configmap "rabbitmq-config-data" not found Nov 24 09:14:21 crc kubenswrapper[4944]: E1124 09:14:21.156903 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:21 crc kubenswrapper[4944]: E1124 09:14:21.156957 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data podName:12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5 nodeName:}" failed. No retries permitted until 2025-11-24 09:14:21.656934085 +0000 UTC m=+1322.191374657 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data") pod "rabbitmq-cell1-server-0" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5") : configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.161025 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-operator-scripts\") pod \"cinder575d-account-delete-8dqqk\" (UID: \"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2\") " pod="openstack/cinder575d-account-delete-8dqqk" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.187127 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-vlctx"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.199839 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z75h\" (UniqueName: \"kubernetes.io/projected/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-kube-api-access-4z75h\") pod \"cinder575d-account-delete-8dqqk\" (UID: \"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2\") " pod="openstack/cinder575d-account-delete-8dqqk" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.206556 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-vlctx"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.292602 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-operator-scripts\") pod \"glance509f-account-delete-fx5bb\" (UID: \"1db23d75-43f2-4bae-9e09-399d0ffb7ed5\") " pod="openstack/glance509f-account-delete-fx5bb" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.292748 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qs7r\" (UniqueName: \"kubernetes.io/projected/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-kube-api-access-5qs7r\") pod \"glance509f-account-delete-fx5bb\" (UID: \"1db23d75-43f2-4bae-9e09-399d0ffb7ed5\") " pod="openstack/glance509f-account-delete-fx5bb" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.305982 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-operator-scripts\") pod \"glance509f-account-delete-fx5bb\" (UID: \"1db23d75-43f2-4bae-9e09-399d0ffb7ed5\") " pod="openstack/glance509f-account-delete-fx5bb" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.315414 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement0cdc-account-delete-sl7j9"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.329767 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement0cdc-account-delete-sl7j9" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.348145 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qs7r\" (UniqueName: \"kubernetes.io/projected/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-kube-api-access-5qs7r\") pod \"glance509f-account-delete-fx5bb\" (UID: \"1db23d75-43f2-4bae-9e09-399d0ffb7ed5\") " pod="openstack/glance509f-account-delete-fx5bb" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.353891 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement0cdc-account-delete-sl7j9"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.376279 4944 generic.go:334] "Generic (PLEG): container finished" podID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerID="81af0f39619d4f0d4399fe088165e6c6fb343c03871f10888e1d787a30b8aa80" exitCode=2 Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.376450 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2dfd2e16-b145-4464-8c5a-bff9a5908d45","Type":"ContainerDied","Data":"81af0f39619d4f0d4399fe088165e6c6fb343c03871f10888e1d787a30b8aa80"} Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.379524 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance509f-account-delete-fx5bb" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.380412 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder575d-account-delete-8dqqk" Nov 24 09:14:21 crc kubenswrapper[4944]: E1124 09:14:21.384769 4944 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-npmfc" message=< Nov 24 09:14:21 crc kubenswrapper[4944]: Exiting ovn-controller (1) [ OK ] Nov 24 09:14:21 crc kubenswrapper[4944]: > Nov 24 09:14:21 crc kubenswrapper[4944]: E1124 09:14:21.384809 4944 kuberuntime_container.go:691] "PreStop hook failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " pod="openstack/ovn-controller-npmfc" podUID="69084f7b-b347-4fdf-917e-2f534f3cc47c" containerName="ovn-controller" containerID="cri-o://b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.384854 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-npmfc" podUID="69084f7b-b347-4fdf-917e-2f534f3cc47c" containerName="ovn-controller" containerID="cri-o://b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009" gracePeriod=30 Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.396608 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69996228-3cfa-486e-8291-12e44f438c5b-operator-scripts\") pod \"placement0cdc-account-delete-sl7j9\" (UID: \"69996228-3cfa-486e-8291-12e44f438c5b\") " pod="openstack/placement0cdc-account-delete-sl7j9" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.396703 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtzf5\" (UniqueName: \"kubernetes.io/projected/69996228-3cfa-486e-8291-12e44f438c5b-kube-api-access-rtzf5\") pod \"placement0cdc-account-delete-sl7j9\" (UID: \"69996228-3cfa-486e-8291-12e44f438c5b\") " pod="openstack/placement0cdc-account-delete-sl7j9" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.500814 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69996228-3cfa-486e-8291-12e44f438c5b-operator-scripts\") pod \"placement0cdc-account-delete-sl7j9\" (UID: \"69996228-3cfa-486e-8291-12e44f438c5b\") " pod="openstack/placement0cdc-account-delete-sl7j9" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.505729 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron7ac6-account-delete-7kth8"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.507243 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtzf5\" (UniqueName: \"kubernetes.io/projected/69996228-3cfa-486e-8291-12e44f438c5b-kube-api-access-rtzf5\") pod \"placement0cdc-account-delete-sl7j9\" (UID: \"69996228-3cfa-486e-8291-12e44f438c5b\") " pod="openstack/placement0cdc-account-delete-sl7j9" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.513178 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69996228-3cfa-486e-8291-12e44f438c5b-operator-scripts\") pod \"placement0cdc-account-delete-sl7j9\" (UID: \"69996228-3cfa-486e-8291-12e44f438c5b\") " pod="openstack/placement0cdc-account-delete-sl7j9" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.527248 4944 generic.go:334] "Generic (PLEG): container finished" podID="5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" containerID="823389b569a8fd3fda96971a913dc73c72a4f283c5b7ab6145f8fb0dd9b6cb9e" exitCode=0 Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.548193 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" event={"ID":"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef","Type":"ContainerDied","Data":"823389b569a8fd3fda96971a913dc73c72a4f283c5b7ab6145f8fb0dd9b6cb9e"} Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.552336 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron7ac6-account-delete-7kth8"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.548833 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron7ac6-account-delete-7kth8" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.556511 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-m2b6x_c521e9a3-0d64-4658-8a00-690d20e619d0/openstack-network-exporter/0.log" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.556557 4944 generic.go:334] "Generic (PLEG): container finished" podID="c521e9a3-0d64-4658-8a00-690d20e619d0" containerID="b41e32588963f3e10325bdeeaf1f1bb9e18dd5fa9618690b49d86fb5f1d1c724" exitCode=2 Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.556593 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-m2b6x" event={"ID":"c521e9a3-0d64-4658-8a00-690d20e619d0","Type":"ContainerDied","Data":"b41e32588963f3e10325bdeeaf1f1bb9e18dd5fa9618690b49d86fb5f1d1c724"} Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.569030 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtzf5\" (UniqueName: \"kubernetes.io/projected/69996228-3cfa-486e-8291-12e44f438c5b-kube-api-access-rtzf5\") pod \"placement0cdc-account-delete-sl7j9\" (UID: \"69996228-3cfa-486e-8291-12e44f438c5b\") " pod="openstack/placement0cdc-account-delete-sl7j9" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.571545 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-s7k5s"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.610180 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-s7k5s"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.612735 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95ckf\" (UniqueName: \"kubernetes.io/projected/11d8c8f9-ee83-4450-9a76-3c0a020652de-kube-api-access-95ckf\") pod \"neutron7ac6-account-delete-7kth8\" (UID: \"11d8c8f9-ee83-4450-9a76-3c0a020652de\") " pod="openstack/neutron7ac6-account-delete-7kth8" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.612872 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d8c8f9-ee83-4450-9a76-3c0a020652de-operator-scripts\") pod \"neutron7ac6-account-delete-7kth8\" (UID: \"11d8c8f9-ee83-4450-9a76-3c0a020652de\") " pod="openstack/neutron7ac6-account-delete-7kth8" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.651891 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapie0a2-account-delete-bvsxd"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.655803 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapie0a2-account-delete-bvsxd" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.677695 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapie0a2-account-delete-bvsxd"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.685143 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.685613 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="2c04aca3-abaa-4c62-8e62-af920276cc50" containerName="openstack-network-exporter" containerID="cri-o://878cbf2bbb403dd3a431aa6956e733ae81fc764310e89c3f03f0d40d6516dc09" gracePeriod=300 Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.699156 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement0cdc-account-delete-sl7j9" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.725190 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95ckf\" (UniqueName: \"kubernetes.io/projected/11d8c8f9-ee83-4450-9a76-3c0a020652de-kube-api-access-95ckf\") pod \"neutron7ac6-account-delete-7kth8\" (UID: \"11d8c8f9-ee83-4450-9a76-3c0a020652de\") " pod="openstack/neutron7ac6-account-delete-7kth8" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.725268 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d8c8f9-ee83-4450-9a76-3c0a020652de-operator-scripts\") pod \"neutron7ac6-account-delete-7kth8\" (UID: \"11d8c8f9-ee83-4450-9a76-3c0a020652de\") " pod="openstack/neutron7ac6-account-delete-7kth8" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.726626 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d8c8f9-ee83-4450-9a76-3c0a020652de-operator-scripts\") pod \"neutron7ac6-account-delete-7kth8\" (UID: \"11d8c8f9-ee83-4450-9a76-3c0a020652de\") " pod="openstack/neutron7ac6-account-delete-7kth8" Nov 24 09:14:21 crc kubenswrapper[4944]: E1124 09:14:21.726906 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:21 crc kubenswrapper[4944]: E1124 09:14:21.726941 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data podName:12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5 nodeName:}" failed. No retries permitted until 2025-11-24 09:14:22.726930407 +0000 UTC m=+1323.261370869 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data") pod "rabbitmq-cell1-server-0" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5") : configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.727183 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.729037 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="b9d45d4c-ab9f-426c-a193-7f46c398ed64" containerName="openstack-network-exporter" containerID="cri-o://3ce79850e6073ff3d7e81da5fd81b8be25dfceb1d21948518006ccbe0b1eda7a" gracePeriod=300 Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.753238 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95ckf\" (UniqueName: \"kubernetes.io/projected/11d8c8f9-ee83-4450-9a76-3c0a020652de-kube-api-access-95ckf\") pod \"neutron7ac6-account-delete-7kth8\" (UID: \"11d8c8f9-ee83-4450-9a76-3c0a020652de\") " pod="openstack/neutron7ac6-account-delete-7kth8" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.779497 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0eec8-account-delete-jtnrx"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.781510 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0eec8-account-delete-jtnrx" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.821871 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0eec8-account-delete-jtnrx"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.826552 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-operator-scripts\") pod \"novaapie0a2-account-delete-bvsxd\" (UID: \"2a65aeee-1ac4-4ae3-b713-ad1d0074f797\") " pod="openstack/novaapie0a2-account-delete-bvsxd" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.826733 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqxb6\" (UniqueName: \"kubernetes.io/projected/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-kube-api-access-jqxb6\") pod \"novaapie0a2-account-delete-bvsxd\" (UID: \"2a65aeee-1ac4-4ae3-b713-ad1d0074f797\") " pod="openstack/novaapie0a2-account-delete-bvsxd" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.891194 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.891774 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="53d02d15-b171-4e78-b528-9bec6362bf70" containerName="cinder-scheduler" containerID="cri-o://b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b" gracePeriod=30 Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.895338 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="53d02d15-b171-4e78-b528-9bec6362bf70" containerName="probe" containerID="cri-o://fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f" gracePeriod=30 Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.898637 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="2c04aca3-abaa-4c62-8e62-af920276cc50" containerName="ovsdbserver-sb" containerID="cri-o://61ee3036c3b98fca180967e52f25b5128d65397ebffba1911bb8a5e25c2ec958" gracePeriod=300 Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.930464 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-operator-scripts\") pod \"novaapie0a2-account-delete-bvsxd\" (UID: \"2a65aeee-1ac4-4ae3-b713-ad1d0074f797\") " pod="openstack/novaapie0a2-account-delete-bvsxd" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.930527 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndpl2\" (UniqueName: \"kubernetes.io/projected/b69c572d-53a7-4d39-a8fc-4152846915a5-kube-api-access-ndpl2\") pod \"novacell0eec8-account-delete-jtnrx\" (UID: \"b69c572d-53a7-4d39-a8fc-4152846915a5\") " pod="openstack/novacell0eec8-account-delete-jtnrx" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.930592 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b69c572d-53a7-4d39-a8fc-4152846915a5-operator-scripts\") pod \"novacell0eec8-account-delete-jtnrx\" (UID: \"b69c572d-53a7-4d39-a8fc-4152846915a5\") " pod="openstack/novacell0eec8-account-delete-jtnrx" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.930624 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqxb6\" (UniqueName: \"kubernetes.io/projected/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-kube-api-access-jqxb6\") pod \"novaapie0a2-account-delete-bvsxd\" (UID: \"2a65aeee-1ac4-4ae3-b713-ad1d0074f797\") " pod="openstack/novaapie0a2-account-delete-bvsxd" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.931676 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-operator-scripts\") pod \"novaapie0a2-account-delete-bvsxd\" (UID: \"2a65aeee-1ac4-4ae3-b713-ad1d0074f797\") " pod="openstack/novaapie0a2-account-delete-bvsxd" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.941496 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-vqngq"] Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.956785 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron7ac6-account-delete-7kth8" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.970664 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqxb6\" (UniqueName: \"kubernetes.io/projected/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-kube-api-access-jqxb6\") pod \"novaapie0a2-account-delete-bvsxd\" (UID: \"2a65aeee-1ac4-4ae3-b713-ad1d0074f797\") " pod="openstack/novaapie0a2-account-delete-bvsxd" Nov 24 09:14:21 crc kubenswrapper[4944]: I1124 09:14:21.979329 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-vqngq"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.000061 4944 log.go:32] "ExecSync cmd from runtime service failed" err=< Nov 24 09:14:22 crc kubenswrapper[4944]: rpc error: code = Unknown desc = command error: setns `mnt`: Bad file descriptor Nov 24 09:14:22 crc kubenswrapper[4944]: fail startup Nov 24 09:14:22 crc kubenswrapper[4944]: , stdout: , stderr: , exit code -1 Nov 24 09:14:22 crc kubenswrapper[4944]: > containerID="61ee3036c3b98fca180967e52f25b5128d65397ebffba1911bb8a5e25c2ec958" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.012294 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 61ee3036c3b98fca180967e52f25b5128d65397ebffba1911bb8a5e25c2ec958 is running failed: container process not found" containerID="61ee3036c3b98fca180967e52f25b5128d65397ebffba1911bb8a5e25c2ec958" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.017161 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 61ee3036c3b98fca180967e52f25b5128d65397ebffba1911bb8a5e25c2ec958 is running failed: container process not found" containerID="61ee3036c3b98fca180967e52f25b5128d65397ebffba1911bb8a5e25c2ec958" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.017225 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 61ee3036c3b98fca180967e52f25b5128d65397ebffba1911bb8a5e25c2ec958 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="2c04aca3-abaa-4c62-8e62-af920276cc50" containerName="ovsdbserver-sb" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.017598 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-mpksr"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.032979 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndpl2\" (UniqueName: \"kubernetes.io/projected/b69c572d-53a7-4d39-a8fc-4152846915a5-kube-api-access-ndpl2\") pod \"novacell0eec8-account-delete-jtnrx\" (UID: \"b69c572d-53a7-4d39-a8fc-4152846915a5\") " pod="openstack/novacell0eec8-account-delete-jtnrx" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.033112 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b69c572d-53a7-4d39-a8fc-4152846915a5-operator-scripts\") pod \"novacell0eec8-account-delete-jtnrx\" (UID: \"b69c572d-53a7-4d39-a8fc-4152846915a5\") " pod="openstack/novacell0eec8-account-delete-jtnrx" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.034380 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b69c572d-53a7-4d39-a8fc-4152846915a5-operator-scripts\") pod \"novacell0eec8-account-delete-jtnrx\" (UID: \"b69c572d-53a7-4d39-a8fc-4152846915a5\") " pod="openstack/novacell0eec8-account-delete-jtnrx" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.051075 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-mpksr"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.051581 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapie0a2-account-delete-bvsxd" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.057641 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="b9d45d4c-ab9f-426c-a193-7f46c398ed64" containerName="ovsdbserver-nb" containerID="cri-o://5ef21e16c9d939c2affb5458115459c6969a680e9f313ed1d7b13c1de4f3f96a" gracePeriod=300 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.073358 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndpl2\" (UniqueName: \"kubernetes.io/projected/b69c572d-53a7-4d39-a8fc-4152846915a5-kube-api-access-ndpl2\") pod \"novacell0eec8-account-delete-jtnrx\" (UID: \"b69c572d-53a7-4d39-a8fc-4152846915a5\") " pod="openstack/novacell0eec8-account-delete-jtnrx" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.104225 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.104742 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b9a8d972-b16f-421e-9fe4-f57653048845" containerName="cinder-api-log" containerID="cri-o://ef2d54ce03a16a4e0d23c0030984dc6c28b947390104ad9f47f988b5741614d5" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.105154 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b9a8d972-b16f-421e-9fe4-f57653048845" containerName="cinder-api" containerID="cri-o://84fadfabdc280facbbd9a3442d721a8b2b1c0298a2c5bd828ac25612a83c8d81" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.128079 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-96smc"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.143659 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovs-vswitchd" containerID="cri-o://c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" gracePeriod=29 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.170546 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-96smc"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.180471 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0eec8-account-delete-jtnrx" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.240616 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-5w7xp"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.242154 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.242320 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data podName:67e92d3f-3532-467f-bf38-c9c3107d4fcb nodeName:}" failed. No retries permitted until 2025-11-24 09:14:24.242304829 +0000 UTC m=+1324.776745281 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data") pod "rabbitmq-server-0" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb") : configmap "rabbitmq-config-data" not found Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.265940 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-5w7xp"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.274578 4944 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 24 09:14:22 crc kubenswrapper[4944]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 24 09:14:22 crc kubenswrapper[4944]: + source /usr/local/bin/container-scripts/functions Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNBridge=br-int Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNRemote=tcp:localhost:6642 Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNEncapType=geneve Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNAvailabilityZones= Nov 24 09:14:22 crc kubenswrapper[4944]: ++ EnableChassisAsGateway=true Nov 24 09:14:22 crc kubenswrapper[4944]: ++ PhysicalNetworks= Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNHostName= Nov 24 09:14:22 crc kubenswrapper[4944]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 24 09:14:22 crc kubenswrapper[4944]: ++ ovs_dir=/var/lib/openvswitch Nov 24 09:14:22 crc kubenswrapper[4944]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 24 09:14:22 crc kubenswrapper[4944]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 24 09:14:22 crc kubenswrapper[4944]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 09:14:22 crc kubenswrapper[4944]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 09:14:22 crc kubenswrapper[4944]: + sleep 0.5 Nov 24 09:14:22 crc kubenswrapper[4944]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 09:14:22 crc kubenswrapper[4944]: + sleep 0.5 Nov 24 09:14:22 crc kubenswrapper[4944]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 09:14:22 crc kubenswrapper[4944]: + cleanup_ovsdb_server_semaphore Nov 24 09:14:22 crc kubenswrapper[4944]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 09:14:22 crc kubenswrapper[4944]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 24 09:14:22 crc kubenswrapper[4944]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-7gtjj" message=< Nov 24 09:14:22 crc kubenswrapper[4944]: Exiting ovsdb-server (5) ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 24 09:14:22 crc kubenswrapper[4944]: + source /usr/local/bin/container-scripts/functions Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNBridge=br-int Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNRemote=tcp:localhost:6642 Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNEncapType=geneve Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNAvailabilityZones= Nov 24 09:14:22 crc kubenswrapper[4944]: ++ EnableChassisAsGateway=true Nov 24 09:14:22 crc kubenswrapper[4944]: ++ PhysicalNetworks= Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNHostName= Nov 24 09:14:22 crc kubenswrapper[4944]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 24 09:14:22 crc kubenswrapper[4944]: ++ ovs_dir=/var/lib/openvswitch Nov 24 09:14:22 crc kubenswrapper[4944]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 24 09:14:22 crc kubenswrapper[4944]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 24 09:14:22 crc kubenswrapper[4944]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 09:14:22 crc kubenswrapper[4944]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 09:14:22 crc kubenswrapper[4944]: + sleep 0.5 Nov 24 09:14:22 crc kubenswrapper[4944]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 09:14:22 crc kubenswrapper[4944]: + sleep 0.5 Nov 24 09:14:22 crc kubenswrapper[4944]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 09:14:22 crc kubenswrapper[4944]: + cleanup_ovsdb_server_semaphore Nov 24 09:14:22 crc kubenswrapper[4944]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 09:14:22 crc kubenswrapper[4944]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 24 09:14:22 crc kubenswrapper[4944]: > Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.274629 4944 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 24 09:14:22 crc kubenswrapper[4944]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 24 09:14:22 crc kubenswrapper[4944]: + source /usr/local/bin/container-scripts/functions Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNBridge=br-int Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNRemote=tcp:localhost:6642 Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNEncapType=geneve Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNAvailabilityZones= Nov 24 09:14:22 crc kubenswrapper[4944]: ++ EnableChassisAsGateway=true Nov 24 09:14:22 crc kubenswrapper[4944]: ++ PhysicalNetworks= Nov 24 09:14:22 crc kubenswrapper[4944]: ++ OVNHostName= Nov 24 09:14:22 crc kubenswrapper[4944]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 24 09:14:22 crc kubenswrapper[4944]: ++ ovs_dir=/var/lib/openvswitch Nov 24 09:14:22 crc kubenswrapper[4944]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 24 09:14:22 crc kubenswrapper[4944]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 24 09:14:22 crc kubenswrapper[4944]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 09:14:22 crc kubenswrapper[4944]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 09:14:22 crc kubenswrapper[4944]: + sleep 0.5 Nov 24 09:14:22 crc kubenswrapper[4944]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 09:14:22 crc kubenswrapper[4944]: + sleep 0.5 Nov 24 09:14:22 crc kubenswrapper[4944]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 24 09:14:22 crc kubenswrapper[4944]: + cleanup_ovsdb_server_semaphore Nov 24 09:14:22 crc kubenswrapper[4944]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 24 09:14:22 crc kubenswrapper[4944]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 24 09:14:22 crc kubenswrapper[4944]: > pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server" containerID="cri-o://35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.274673 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server" containerID="cri-o://35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" gracePeriod=29 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.275148 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5ccb7b4f9b-mcdmd"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.275411 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5ccb7b4f9b-mcdmd" podUID="70a59d46-95bf-4681-b61d-22239c638737" containerName="placement-log" containerID="cri-o://4ad8bef4c684b16347a1d2ba8be599c596216230dd5dd3d9c7c090945c16c531" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.275880 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5ccb7b4f9b-mcdmd" podUID="70a59d46-95bf-4681-b61d-22239c638737" containerName="placement-api" containerID="cri-o://e42a48dc5f9a748005eb91765bdfa38b2ad4b4f974b03b15a72a6b00023e47a2" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.307726 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.310117 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.310328 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.310425 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009 is running failed: container process not found" containerID="b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.312207 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009 is running failed: container process not found" containerID="b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.315722 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.315793 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server" Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.316232 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009 is running failed: container process not found" containerID="b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.316275 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-npmfc" podUID="69084f7b-b347-4fdf-917e-2f534f3cc47c" containerName="ovn-controller" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.329745 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0493ffa7-e203-4f40-b117-07b07fe13380" path="/var/lib/kubelet/pods/0493ffa7-e203-4f40-b117-07b07fe13380/volumes" Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.330913 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.331578 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="587209cc-b2bb-460e-a6d4-1391b6294864" path="/var/lib/kubelet/pods/587209cc-b2bb-460e-a6d4-1391b6294864/volumes" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.332384 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b3396e1-991a-4a6e-a7a6-665658af7e63" path="/var/lib/kubelet/pods/6b3396e1-991a-4a6e-a7a6-665658af7e63/volumes" Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.333836 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.333909 4944 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovs-vswitchd" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.333566 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c338ca6-2795-4c94-b863-23579147b584" path="/var/lib/kubelet/pods/6c338ca6-2795-4c94-b863-23579147b584/volumes" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.335153 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="836f8fae-73e8-423e-8dec-0ee779d09985" path="/var/lib/kubelet/pods/836f8fae-73e8-423e-8dec-0ee779d09985/volumes" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.335619 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e780a24-2c73-45db-89c0-00a4ec351220" path="/var/lib/kubelet/pods/8e780a24-2c73-45db-89c0-00a4ec351220/volumes" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.336150 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93" path="/var/lib/kubelet/pods/b4e37c3c-32ab-45ff-8b5a-a46fa5fe0d93/volumes" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.338019 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5349072-d740-47d5-a667-05670688b5c5" path="/var/lib/kubelet/pods/e5349072-d740-47d5-a667-05670688b5c5/volumes" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.338680 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.338719 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.338738 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.338752 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7b47cdb5f8-g5n25"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.339942 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f94b9960-fa37-49f7-9fa7-0e01f855992b" containerName="glance-log" containerID="cri-o://95d6f710f056de96375b740c77544fefc875b760b049fb6331a5d9ea4510ffb4" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.340440 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f94b9960-fa37-49f7-9fa7-0e01f855992b" containerName="glance-httpd" containerID="cri-o://9a5211b2a8ad4ed0c93e5f29ec422b4395da812067c859ba887bf13db9c4afb6" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.348454 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-server" containerID="cri-o://3d31ceee24450aa9fe538636500f7868bbd37f5da8e0012f05ca3b457f8720eb" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.348801 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="swift-recon-cron" containerID="cri-o://d6d5476d01856b9678b39efa998de2c30d0cc6772d172d0bf1aa500eb8ebaca7" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.348840 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="rsync" containerID="cri-o://be411b5a8274b541c609635739f60faee6efa6abf3fbf8c48eed07bef8944385" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.348870 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-expirer" containerID="cri-o://48d0fed6b1c1123b8f3c416f72272d6749327411fdfbf987f7ee43dd423e071e" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.348904 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-updater" containerID="cri-o://6ccdc9b46a378806550b2c6500f4fd45536785660f1ba53f9c389a60f6a9bf41" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.348935 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-auditor" containerID="cri-o://15e66884ade4e39577226708f15ca7622732f6b82d81fb79dff82cc2578f9a9a" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.348963 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-replicator" containerID="cri-o://e6c6e71ef8c3c8177a39c99dabb96e105d42c681788bccbca30c6f9504cba707" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.348993 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-server" containerID="cri-o://bb4e495759495e1efa8e3c7bf1b539d36a09c8a14b177fd703baec61783412a9" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.349021 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-updater" containerID="cri-o://c17e6f2fe0c81b82e5511368c8268b7225365160cc97761f7dc0ee157f7a8484" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.349113 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-auditor" containerID="cri-o://91b30be5efcaf3b6adec37e7c1e859fed2c520997d002c00b9500d39d521b0ed" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.349143 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-replicator" containerID="cri-o://e533818c9ecf15f88341998771d0639c9cac4cfb451d5374ad40bec2e815a3ff" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.349170 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-server" containerID="cri-o://5d2fbee82caba7d0ca08ca333bcbf3e804351883721f48d8d9bfaac47a78f960" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.349199 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-reaper" containerID="cri-o://87e895d554127f520ab155e47e47d2faa5077bd4196802ecc68ee9b04403fe3d" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.349270 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-auditor" containerID="cri-o://9311cf4164c78aa256b1228bc545098e231c1036813339b9399c8c0ccdf8463f" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.349336 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" podUID="e3ea29cb-ae71-48af-8005-ba99c178a573" containerName="barbican-keystone-listener" containerID="cri-o://e3db2c95e6254ee19d5cd78241174e4c4f6100bc9626f271af3b94acb76058e6" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.349323 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-replicator" containerID="cri-o://2842b8bb44b04f4fb4de40704e3d22283e14c59ce261e4a211bef21dbb6e7a96" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.351070 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" podUID="e3ea29cb-ae71-48af-8005-ba99c178a573" containerName="barbican-keystone-listener-log" containerID="cri-o://5e2bf771bd1be4b28e3fa735cec21fcbbccd1e4807f6ac424dd4405dc8f8a329" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.395559 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.395797 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7209b8a6-5a97-4611-8183-84f1db25e2d1" containerName="glance-log" containerID="cri-o://b48283673309355d5ff56fb2136836b9670187973eb44ca2cc22b5681057bccf" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.396180 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7209b8a6-5a97-4611-8183-84f1db25e2d1" containerName="glance-httpd" containerID="cri-o://71ca2485422737d39ab2f15fc0ad24cd930f9cf353b7e908277d85b6a71d3f14" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.441186 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="67e92d3f-3532-467f-bf38-c9c3107d4fcb" containerName="rabbitmq" containerID="cri-o://ea9e352d0f24eca87b4c56a85b5935690423ccc0712e37d2ea78fde15ddd4336" gracePeriod=604800 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.454915 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-58f6cff88f-s86kk"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.455394 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-58f6cff88f-s86kk" podUID="3d9c665f-2be7-4b08-934d-b63036656b5f" containerName="barbican-worker-log" containerID="cri-o://219c63bcbbda1c8ee292c94dcc9e2f39b33c5aab0d972baaee0d1dc1f454692d" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.457882 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-58f6cff88f-s86kk" podUID="3d9c665f-2be7-4b08-934d-b63036656b5f" containerName="barbican-worker" containerID="cri-o://9906021955dc6929c0ac49f16b105afc15f33b120aab600fc31ecf2026e5ba6e" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.664150 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d5d955fb7-kf4q4"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.664704 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d5d955fb7-kf4q4" podUID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" containerName="neutron-api" containerID="cri-o://ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.672705 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d5d955fb7-kf4q4" podUID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" containerName="neutron-httpd" containerID="cri-o://dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.675794 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f49db88bb-dzrb6"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.676039 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f49db88bb-dzrb6" podUID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerName="barbican-api-log" containerID="cri-o://7aa8ffc64bce11fa46956c14a0b6ed75513f2964c1dd926a5e6be5120e48e88e" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.676188 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f49db88bb-dzrb6" podUID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerName="barbican-api" containerID="cri-o://7159a46717426d269e3e9ca766c99432598e2da70a9cf82e6a43abda53ec27b9" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.683734 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.690300 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.690493 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-log" containerID="cri-o://3dfe8ae924d36e9fb2dbf24bc72c97e4e845aaa92a56c64d0448011249204cfb" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.690619 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-metadata" containerID="cri-o://f4fda6a5047a5412abc68148082d7806e36494cd3997be68245586f663ec9996" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.699148 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.699498 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerName="nova-api-log" containerID="cri-o://ceef1855e41bb1981a2f6e513be92fa9f3b4e90878c9354637a799ca0b3bbbf4" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.699606 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerName="nova-api-api" containerID="cri-o://1156a54f88c1c2dd0577bc2b082f7256b862e090c82e8172fc3df6b5bf9e5bdf" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.706690 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-d8m52"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.714164 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-2d40-account-create-l7jv5"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.745117 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-d8m52"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.754421 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-2d40-account-create-l7jv5"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.764076 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.764360 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="709b6d2b-7e0f-4f60-9bd2-edacb409a727" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://bc869a248132ca4279a0e2f68eec18b02bc447d39b21ce0e334fb974d4d68487" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.774214 4944 generic.go:334] "Generic (PLEG): container finished" podID="70a59d46-95bf-4681-b61d-22239c638737" containerID="4ad8bef4c684b16347a1d2ba8be599c596216230dd5dd3d9c7c090945c16c531" exitCode=143 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.774279 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5ccb7b4f9b-mcdmd" event={"ID":"70a59d46-95bf-4681-b61d-22239c638737","Type":"ContainerDied","Data":"4ad8bef4c684b16347a1d2ba8be599c596216230dd5dd3d9c7c090945c16c531"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.780696 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.781740 4944 generic.go:334] "Generic (PLEG): container finished" podID="69084f7b-b347-4fdf-917e-2f534f3cc47c" containerID="b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009" exitCode=0 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.781797 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-npmfc" event={"ID":"69084f7b-b347-4fdf-917e-2f534f3cc47c","Type":"ContainerDied","Data":"b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.785188 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b9d45d4c-ab9f-426c-a193-7f46c398ed64/ovsdbserver-nb/0.log" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.785265 4944 generic.go:334] "Generic (PLEG): container finished" podID="b9d45d4c-ab9f-426c-a193-7f46c398ed64" containerID="3ce79850e6073ff3d7e81da5fd81b8be25dfceb1d21948518006ccbe0b1eda7a" exitCode=2 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.785285 4944 generic.go:334] "Generic (PLEG): container finished" podID="b9d45d4c-ab9f-426c-a193-7f46c398ed64" containerID="5ef21e16c9d939c2affb5458115459c6969a680e9f313ed1d7b13c1de4f3f96a" exitCode=143 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.786931 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b9d45d4c-ab9f-426c-a193-7f46c398ed64","Type":"ContainerDied","Data":"3ce79850e6073ff3d7e81da5fd81b8be25dfceb1d21948518006ccbe0b1eda7a"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.786965 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b9d45d4c-ab9f-426c-a193-7f46c398ed64","Type":"ContainerDied","Data":"5ef21e16c9d939c2affb5458115459c6969a680e9f313ed1d7b13c1de4f3f96a"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.786980 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.789453 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="90fb3cc2-0725-48dc-9ad5-5c8da27930f8" containerName="nova-scheduler-scheduler" containerID="cri-o://58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.793062 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:22 crc kubenswrapper[4944]: E1124 09:14:22.793135 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data podName:12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5 nodeName:}" failed. No retries permitted until 2025-11-24 09:14:24.793113643 +0000 UTC m=+1325.327554105 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data") pod "rabbitmq-cell1-server-0" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5") : configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.800190 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-m2b6x_c521e9a3-0d64-4658-8a00-690d20e619d0/openstack-network-exporter/0.log" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.800264 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.821663 4944 generic.go:334] "Generic (PLEG): container finished" podID="b9a8d972-b16f-421e-9fe4-f57653048845" containerID="ef2d54ce03a16a4e0d23c0030984dc6c28b947390104ad9f47f988b5741614d5" exitCode=143 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.821740 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a8d972-b16f-421e-9fe4-f57653048845","Type":"ContainerDied","Data":"ef2d54ce03a16a4e0d23c0030984dc6c28b947390104ad9f47f988b5741614d5"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.835135 4944 generic.go:334] "Generic (PLEG): container finished" podID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" exitCode=0 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.835605 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7gtjj" event={"ID":"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b","Type":"ContainerDied","Data":"35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.858532 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2c04aca3-abaa-4c62-8e62-af920276cc50/ovsdbserver-sb/0.log" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.858590 4944 generic.go:334] "Generic (PLEG): container finished" podID="2c04aca3-abaa-4c62-8e62-af920276cc50" containerID="878cbf2bbb403dd3a431aa6956e733ae81fc764310e89c3f03f0d40d6516dc09" exitCode=2 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.858609 4944 generic.go:334] "Generic (PLEG): container finished" podID="2c04aca3-abaa-4c62-8e62-af920276cc50" containerID="61ee3036c3b98fca180967e52f25b5128d65397ebffba1911bb8a5e25c2ec958" exitCode=143 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.858661 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2c04aca3-abaa-4c62-8e62-af920276cc50","Type":"ContainerDied","Data":"878cbf2bbb403dd3a431aa6956e733ae81fc764310e89c3f03f0d40d6516dc09"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.858693 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2c04aca3-abaa-4c62-8e62-af920276cc50","Type":"ContainerDied","Data":"61ee3036c3b98fca180967e52f25b5128d65397ebffba1911bb8a5e25c2ec958"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.875479 4944 generic.go:334] "Generic (PLEG): container finished" podID="e3ea29cb-ae71-48af-8005-ba99c178a573" containerID="5e2bf771bd1be4b28e3fa735cec21fcbbccd1e4807f6ac424dd4405dc8f8a329" exitCode=143 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.875552 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" event={"ID":"e3ea29cb-ae71-48af-8005-ba99c178a573","Type":"ContainerDied","Data":"5e2bf771bd1be4b28e3fa735cec21fcbbccd1e4807f6ac424dd4405dc8f8a329"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.877704 4944 generic.go:334] "Generic (PLEG): container finished" podID="f94b9960-fa37-49f7-9fa7-0e01f855992b" containerID="95d6f710f056de96375b740c77544fefc875b760b049fb6331a5d9ea4510ffb4" exitCode=143 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.877751 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f94b9960-fa37-49f7-9fa7-0e01f855992b","Type":"ContainerDied","Data":"95d6f710f056de96375b740c77544fefc875b760b049fb6331a5d9ea4510ffb4"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.880831 4944 generic.go:334] "Generic (PLEG): container finished" podID="7209b8a6-5a97-4611-8183-84f1db25e2d1" containerID="b48283673309355d5ff56fb2136836b9670187973eb44ca2cc22b5681057bccf" exitCode=143 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.880919 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7209b8a6-5a97-4611-8183-84f1db25e2d1","Type":"ContainerDied","Data":"b48283673309355d5ff56fb2136836b9670187973eb44ca2cc22b5681057bccf"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.893691 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-combined-ca-bundle\") pod \"c521e9a3-0d64-4658-8a00-690d20e619d0\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.893762 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grfb9\" (UniqueName: \"kubernetes.io/projected/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-kube-api-access-grfb9\") pod \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.893794 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-metrics-certs-tls-certs\") pod \"c521e9a3-0d64-4658-8a00-690d20e619d0\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.893888 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c521e9a3-0d64-4658-8a00-690d20e619d0-config\") pod \"c521e9a3-0d64-4658-8a00-690d20e619d0\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.893909 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-config\") pod \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.893949 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcfxw\" (UniqueName: \"kubernetes.io/projected/c521e9a3-0d64-4658-8a00-690d20e619d0-kube-api-access-hcfxw\") pod \"c521e9a3-0d64-4658-8a00-690d20e619d0\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.894008 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-svc\") pod \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.894036 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-nb\") pod \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.894089 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-swift-storage-0\") pod \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.894139 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-sb\") pod \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\" (UID: \"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.894163 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovs-rundir\") pod \"c521e9a3-0d64-4658-8a00-690d20e619d0\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.894192 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovn-rundir\") pod \"c521e9a3-0d64-4658-8a00-690d20e619d0\" (UID: \"c521e9a3-0d64-4658-8a00-690d20e619d0\") " Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.894621 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "c521e9a3-0d64-4658-8a00-690d20e619d0" (UID: "c521e9a3-0d64-4658-8a00-690d20e619d0"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.897724 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-p2c5n"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.916188 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "c521e9a3-0d64-4658-8a00-690d20e619d0" (UID: "c521e9a3-0d64-4658-8a00-690d20e619d0"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.916774 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c521e9a3-0d64-4658-8a00-690d20e619d0-config" (OuterVolumeSpecName: "config") pod "c521e9a3-0d64-4658-8a00-690d20e619d0" (UID: "c521e9a3-0d64-4658-8a00-690d20e619d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.916805 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.916997 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32" containerName="nova-cell1-conductor-conductor" containerID="cri-o://83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b" gracePeriod=30 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952433 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="15e66884ade4e39577226708f15ca7622732f6b82d81fb79dff82cc2578f9a9a" exitCode=0 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952468 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="c17e6f2fe0c81b82e5511368c8268b7225365160cc97761f7dc0ee157f7a8484" exitCode=0 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952477 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="91b30be5efcaf3b6adec37e7c1e859fed2c520997d002c00b9500d39d521b0ed" exitCode=0 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952485 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="e533818c9ecf15f88341998771d0639c9cac4cfb451d5374ad40bec2e815a3ff" exitCode=0 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952493 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="87e895d554127f520ab155e47e47d2faa5077bd4196802ecc68ee9b04403fe3d" exitCode=0 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952501 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="2842b8bb44b04f4fb4de40704e3d22283e14c59ce261e4a211bef21dbb6e7a96" exitCode=0 Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952527 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"15e66884ade4e39577226708f15ca7622732f6b82d81fb79dff82cc2578f9a9a"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952557 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"c17e6f2fe0c81b82e5511368c8268b7225365160cc97761f7dc0ee157f7a8484"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952570 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"91b30be5efcaf3b6adec37e7c1e859fed2c520997d002c00b9500d39d521b0ed"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952581 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"e533818c9ecf15f88341998771d0639c9cac4cfb451d5374ad40bec2e815a3ff"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952590 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"87e895d554127f520ab155e47e47d2faa5077bd4196802ecc68ee9b04403fe3d"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.952599 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"2842b8bb44b04f4fb4de40704e3d22283e14c59ce261e4a211bef21dbb6e7a96"} Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.959874 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-p2c5n"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.969877 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-kube-api-access-grfb9" (OuterVolumeSpecName: "kube-api-access-grfb9") pod "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" (UID: "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef"). InnerVolumeSpecName "kube-api-access-grfb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.971416 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c521e9a3-0d64-4658-8a00-690d20e619d0-kube-api-access-hcfxw" (OuterVolumeSpecName: "kube-api-access-hcfxw") pod "c521e9a3-0d64-4658-8a00-690d20e619d0" (UID: "c521e9a3-0d64-4658-8a00-690d20e619d0"). InnerVolumeSpecName "kube-api-access-hcfxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.980567 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ttlw6"] Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.998735 4944 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.998782 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grfb9\" (UniqueName: \"kubernetes.io/projected/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-kube-api-access-grfb9\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.998823 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c521e9a3-0d64-4658-8a00-690d20e619d0-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.998836 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcfxw\" (UniqueName: \"kubernetes.io/projected/c521e9a3-0d64-4658-8a00-690d20e619d0-kube-api-access-hcfxw\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:22 crc kubenswrapper[4944]: I1124 09:14:22.998850 4944 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c521e9a3-0d64-4658-8a00-690d20e619d0-ovs-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.002607 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.004403 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c521e9a3-0d64-4658-8a00-690d20e619d0" (UID: "c521e9a3-0d64-4658-8a00-690d20e619d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.008059 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="a8d71630-681e-4126-bb64-e8b03f83ce97" containerName="nova-cell0-conductor-conductor" containerID="cri-o://2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef" gracePeriod=30 Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.057484 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ttlw6"] Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.065293 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" (UID: "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.085302 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-config" (OuterVolumeSpecName: "config") pod "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" (UID: "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.097393 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicand270-account-delete-h2tcc"] Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.101175 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.101554 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.101763 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.131010 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" (UID: "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.167483 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="e454a738-0872-41d8-9432-17432276248c" containerName="galera" containerID="cri-o://06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d" gracePeriod=30 Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.173832 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" (UID: "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.201195 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "c521e9a3-0d64-4658-8a00-690d20e619d0" (UID: "c521e9a3-0d64-4658-8a00-690d20e619d0"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.204977 4944 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c521e9a3-0d64-4658-8a00-690d20e619d0-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.205013 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.205024 4944 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.270310 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" (UID: "5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.308637 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.409310 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.512790 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrvn4\" (UniqueName: \"kubernetes.io/projected/69084f7b-b347-4fdf-917e-2f534f3cc47c-kube-api-access-qrvn4\") pod \"69084f7b-b347-4fdf-917e-2f534f3cc47c\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.513208 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run\") pod \"69084f7b-b347-4fdf-917e-2f534f3cc47c\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.513252 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-ovn-controller-tls-certs\") pod \"69084f7b-b347-4fdf-917e-2f534f3cc47c\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.513363 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-log-ovn\") pod \"69084f7b-b347-4fdf-917e-2f534f3cc47c\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.513377 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run" (OuterVolumeSpecName: "var-run") pod "69084f7b-b347-4fdf-917e-2f534f3cc47c" (UID: "69084f7b-b347-4fdf-917e-2f534f3cc47c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.513401 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-combined-ca-bundle\") pod \"69084f7b-b347-4fdf-917e-2f534f3cc47c\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.513439 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run-ovn\") pod \"69084f7b-b347-4fdf-917e-2f534f3cc47c\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.513582 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69084f7b-b347-4fdf-917e-2f534f3cc47c-scripts\") pod \"69084f7b-b347-4fdf-917e-2f534f3cc47c\" (UID: \"69084f7b-b347-4fdf-917e-2f534f3cc47c\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.514169 4944 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.516084 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69084f7b-b347-4fdf-917e-2f534f3cc47c-scripts" (OuterVolumeSpecName: "scripts") pod "69084f7b-b347-4fdf-917e-2f534f3cc47c" (UID: "69084f7b-b347-4fdf-917e-2f534f3cc47c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.516158 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "69084f7b-b347-4fdf-917e-2f534f3cc47c" (UID: "69084f7b-b347-4fdf-917e-2f534f3cc47c"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.516914 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "69084f7b-b347-4fdf-917e-2f534f3cc47c" (UID: "69084f7b-b347-4fdf-917e-2f534f3cc47c"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.520164 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69084f7b-b347-4fdf-917e-2f534f3cc47c-kube-api-access-qrvn4" (OuterVolumeSpecName: "kube-api-access-qrvn4") pod "69084f7b-b347-4fdf-917e-2f534f3cc47c" (UID: "69084f7b-b347-4fdf-917e-2f534f3cc47c"). InnerVolumeSpecName "kube-api-access-qrvn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.522091 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7bcd854fd7-z2d6k"] Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.527943 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" podUID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerName="proxy-server" containerID="cri-o://2c1a579178840a65cb60ad40f3b83a1c327413e6a2d77e8178e33176b65db789" gracePeriod=30 Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.530632 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" podUID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerName="proxy-httpd" containerID="cri-o://b626ffc5d738486bd96a728516fb7e89c1bf6800fb2f82fd0b1bc9e2f6d19268" gracePeriod=30 Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.548512 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.548567 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.577344 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2c04aca3-abaa-4c62-8e62-af920276cc50/ovsdbserver-sb/0.log" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.577435 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.578220 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b9d45d4c-ab9f-426c-a193-7f46c398ed64/ovsdbserver-nb/0.log" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.578343 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.615319 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69084f7b-b347-4fdf-917e-2f534f3cc47c" (UID: "69084f7b-b347-4fdf-917e-2f534f3cc47c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.616642 4944 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.616666 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69084f7b-b347-4fdf-917e-2f534f3cc47c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.616677 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrvn4\" (UniqueName: \"kubernetes.io/projected/69084f7b-b347-4fdf-917e-2f534f3cc47c-kube-api-access-qrvn4\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.616688 4944 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/69084f7b-b347-4fdf-917e-2f534f3cc47c-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.616698 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.678884 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder575d-account-delete-8dqqk"] Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718479 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4xgf\" (UniqueName: \"kubernetes.io/projected/2c04aca3-abaa-4c62-8e62-af920276cc50-kube-api-access-r4xgf\") pod \"2c04aca3-abaa-4c62-8e62-af920276cc50\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718522 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdb-rundir\") pod \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718542 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdbserver-sb-tls-certs\") pod \"2c04aca3-abaa-4c62-8e62-af920276cc50\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718610 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-combined-ca-bundle\") pod \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718685 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdbserver-nb-tls-certs\") pod \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718728 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-metrics-certs-tls-certs\") pod \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718751 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-scripts\") pod \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718775 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h4qw\" (UniqueName: \"kubernetes.io/projected/b9d45d4c-ab9f-426c-a193-7f46c398ed64-kube-api-access-4h4qw\") pod \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718791 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-scripts\") pod \"2c04aca3-abaa-4c62-8e62-af920276cc50\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718811 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdb-rundir\") pod \"2c04aca3-abaa-4c62-8e62-af920276cc50\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718831 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-config\") pod \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718852 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\" (UID: \"b9d45d4c-ab9f-426c-a193-7f46c398ed64\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718874 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-metrics-certs-tls-certs\") pod \"2c04aca3-abaa-4c62-8e62-af920276cc50\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.718894 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-config\") pod \"2c04aca3-abaa-4c62-8e62-af920276cc50\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.719060 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-combined-ca-bundle\") pod \"2c04aca3-abaa-4c62-8e62-af920276cc50\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.719095 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"2c04aca3-abaa-4c62-8e62-af920276cc50\" (UID: \"2c04aca3-abaa-4c62-8e62-af920276cc50\") " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.723427 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "b9d45d4c-ab9f-426c-a193-7f46c398ed64" (UID: "b9d45d4c-ab9f-426c-a193-7f46c398ed64"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.725437 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-scripts" (OuterVolumeSpecName: "scripts") pod "b9d45d4c-ab9f-426c-a193-7f46c398ed64" (UID: "b9d45d4c-ab9f-426c-a193-7f46c398ed64"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.725607 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "2c04aca3-abaa-4c62-8e62-af920276cc50" (UID: "2c04aca3-abaa-4c62-8e62-af920276cc50"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.725748 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-config" (OuterVolumeSpecName: "config") pod "b9d45d4c-ab9f-426c-a193-7f46c398ed64" (UID: "b9d45d4c-ab9f-426c-a193-7f46c398ed64"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.725795 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-config" (OuterVolumeSpecName: "config") pod "2c04aca3-abaa-4c62-8e62-af920276cc50" (UID: "2c04aca3-abaa-4c62-8e62-af920276cc50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.726017 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-scripts" (OuterVolumeSpecName: "scripts") pod "2c04aca3-abaa-4c62-8e62-af920276cc50" (UID: "2c04aca3-abaa-4c62-8e62-af920276cc50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.734243 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "69084f7b-b347-4fdf-917e-2f534f3cc47c" (UID: "69084f7b-b347-4fdf-917e-2f534f3cc47c"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.735199 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9d45d4c-ab9f-426c-a193-7f46c398ed64-kube-api-access-4h4qw" (OuterVolumeSpecName: "kube-api-access-4h4qw") pod "b9d45d4c-ab9f-426c-a193-7f46c398ed64" (UID: "b9d45d4c-ab9f-426c-a193-7f46c398ed64"). InnerVolumeSpecName "kube-api-access-4h4qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.735274 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "2c04aca3-abaa-4c62-8e62-af920276cc50" (UID: "2c04aca3-abaa-4c62-8e62-af920276cc50"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.752072 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement0cdc-account-delete-sl7j9"] Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.770189 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c04aca3-abaa-4c62-8e62-af920276cc50-kube-api-access-r4xgf" (OuterVolumeSpecName: "kube-api-access-r4xgf") pod "2c04aca3-abaa-4c62-8e62-af920276cc50" (UID: "2c04aca3-abaa-4c62-8e62-af920276cc50"). InnerVolumeSpecName "kube-api-access-r4xgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.777348 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "b9d45d4c-ab9f-426c-a193-7f46c398ed64" (UID: "b9d45d4c-ab9f-426c-a193-7f46c398ed64"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821169 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9d45d4c-ab9f-426c-a193-7f46c398ed64" (UID: "b9d45d4c-ab9f-426c-a193-7f46c398ed64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821545 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821584 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4xgf\" (UniqueName: \"kubernetes.io/projected/2c04aca3-abaa-4c62-8e62-af920276cc50-kube-api-access-r4xgf\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821599 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821611 4944 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/69084f7b-b347-4fdf-917e-2f534f3cc47c-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821625 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821636 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821647 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h4qw\" (UniqueName: \"kubernetes.io/projected/b9d45d4c-ab9f-426c-a193-7f46c398ed64-kube-api-access-4h4qw\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821660 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821670 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821681 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9d45d4c-ab9f-426c-a193-7f46c398ed64-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821727 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.821742 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c04aca3-abaa-4c62-8e62-af920276cc50-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.855906 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.918157 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c04aca3-abaa-4c62-8e62-af920276cc50" (UID: "2c04aca3-abaa-4c62-8e62-af920276cc50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:23 crc kubenswrapper[4944]: I1124 09:14:23.939785 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.085407 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.087886 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.087933 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.087944 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.088086 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "2c04aca3-abaa-4c62-8e62-af920276cc50" (UID: "2c04aca3-abaa-4c62-8e62-af920276cc50"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.097290 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "2c04aca3-abaa-4c62-8e62-af920276cc50" (UID: "2c04aca3-abaa-4c62-8e62-af920276cc50"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.099435 4944 generic.go:334] "Generic (PLEG): container finished" podID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" containerID="dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.099493 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d5d955fb7-kf4q4" event={"ID":"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6","Type":"ContainerDied","Data":"dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.100581 4944 generic.go:334] "Generic (PLEG): container finished" podID="709b6d2b-7e0f-4f60-9bd2-edacb409a727" containerID="bc869a248132ca4279a0e2f68eec18b02bc447d39b21ce0e334fb974d4d68487" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.100618 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"709b6d2b-7e0f-4f60-9bd2-edacb409a727","Type":"ContainerDied","Data":"bc869a248132ca4279a0e2f68eec18b02bc447d39b21ce0e334fb974d4d68487"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111037 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="be411b5a8274b541c609635739f60faee6efa6abf3fbf8c48eed07bef8944385" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111099 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="48d0fed6b1c1123b8f3c416f72272d6749327411fdfbf987f7ee43dd423e071e" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111111 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="6ccdc9b46a378806550b2c6500f4fd45536785660f1ba53f9c389a60f6a9bf41" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111120 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="e6c6e71ef8c3c8177a39c99dabb96e105d42c681788bccbca30c6f9504cba707" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111146 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="bb4e495759495e1efa8e3c7bf1b539d36a09c8a14b177fd703baec61783412a9" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111155 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="5d2fbee82caba7d0ca08ca333bcbf3e804351883721f48d8d9bfaac47a78f960" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111164 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="9311cf4164c78aa256b1228bc545098e231c1036813339b9399c8c0ccdf8463f" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111171 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="3d31ceee24450aa9fe538636500f7868bbd37f5da8e0012f05ca3b457f8720eb" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111531 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"be411b5a8274b541c609635739f60faee6efa6abf3fbf8c48eed07bef8944385"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111588 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"48d0fed6b1c1123b8f3c416f72272d6749327411fdfbf987f7ee43dd423e071e"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111605 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"6ccdc9b46a378806550b2c6500f4fd45536785660f1ba53f9c389a60f6a9bf41"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111617 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"e6c6e71ef8c3c8177a39c99dabb96e105d42c681788bccbca30c6f9504cba707"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111629 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"bb4e495759495e1efa8e3c7bf1b539d36a09c8a14b177fd703baec61783412a9"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111639 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"5d2fbee82caba7d0ca08ca333bcbf3e804351883721f48d8d9bfaac47a78f960"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111667 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"9311cf4164c78aa256b1228bc545098e231c1036813339b9399c8c0ccdf8463f"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.111679 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"3d31ceee24450aa9fe538636500f7868bbd37f5da8e0012f05ca3b457f8720eb"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.118276 4944 generic.go:334] "Generic (PLEG): container finished" podID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerID="3dfe8ae924d36e9fb2dbf24bc72c97e4e845aaa92a56c64d0448011249204cfb" exitCode=143 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.118393 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc12ffc5-e026-42a6-8ddc-43961c535787","Type":"ContainerDied","Data":"3dfe8ae924d36e9fb2dbf24bc72c97e4e845aaa92a56c64d0448011249204cfb"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.124574 4944 generic.go:334] "Generic (PLEG): container finished" podID="53d02d15-b171-4e78-b528-9bec6362bf70" containerID="fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.124690 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"53d02d15-b171-4e78-b528-9bec6362bf70","Type":"ContainerDied","Data":"fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.126849 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "b9d45d4c-ab9f-426c-a193-7f46c398ed64" (UID: "b9d45d4c-ab9f-426c-a193-7f46c398ed64"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.131511 4944 generic.go:334] "Generic (PLEG): container finished" podID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerID="7aa8ffc64bce11fa46956c14a0b6ed75513f2964c1dd926a5e6be5120e48e88e" exitCode=143 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.131639 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f49db88bb-dzrb6" event={"ID":"046a8c15-6acd-47fe-aef7-daac5085c2ec","Type":"ContainerDied","Data":"7aa8ffc64bce11fa46956c14a0b6ed75513f2964c1dd926a5e6be5120e48e88e"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.142331 4944 generic.go:334] "Generic (PLEG): container finished" podID="9a8dab89-ec08-4497-a890-bcb70c8ea430" containerID="194effb1be878a12474f5aad321c9b034378e690429268d01ba3188c16709667" exitCode=137 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.142440 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0888fb3045d88281da186d7e257c308ceb1dccdcf74fe524fc0b7cdbc1258ec" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.150569 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-npmfc" event={"ID":"69084f7b-b347-4fdf-917e-2f534f3cc47c","Type":"ContainerDied","Data":"9f239ec933ea933cd803c9ec051a6a8fa85566e266369c4d911776e1e04db753"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.150617 4944 scope.go:117] "RemoveContainer" containerID="b0356ed9cac9eca4ffb85e1befcff9b8280973a79aceb2bdaadc68bfcba87009" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.150747 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-npmfc" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.162107 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" event={"ID":"5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef","Type":"ContainerDied","Data":"155986593ce3bec854130e84ea3700c6931f55e508517cb83b1791e49d0596ed"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.162214 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.169608 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "b9d45d4c-ab9f-426c-a193-7f46c398ed64" (UID: "b9d45d4c-ab9f-426c-a193-7f46c398ed64"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.180669 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicand270-account-delete-h2tcc" event={"ID":"55a9098c-7257-46f7-888c-c4bb6f9421a5","Type":"ContainerStarted","Data":"f79b603137a9e552f12da861f283799163e55db755214ecf4ce6c275a2f775c7"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.180710 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicand270-account-delete-h2tcc" event={"ID":"55a9098c-7257-46f7-888c-c4bb6f9421a5","Type":"ContainerStarted","Data":"18abcd5217b86e8d889c6cc23c83d4771c98be7cd169a305a78158c2bd5ad07b"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.183610 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-m2b6x_c521e9a3-0d64-4658-8a00-690d20e619d0/openstack-network-exporter/0.log" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.183777 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-m2b6x" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.184432 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-m2b6x" event={"ID":"c521e9a3-0d64-4658-8a00-690d20e619d0","Type":"ContainerDied","Data":"98bdfdff50d3326231df5a40e90f31540c340f817eb18f1f93ed08e357d3228f"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.189415 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.189654 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.189747 4944 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9d45d4c-ab9f-426c-a193-7f46c398ed64-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.189877 4944 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c04aca3-abaa-4c62-8e62-af920276cc50-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.197556 4944 generic.go:334] "Generic (PLEG): container finished" podID="3d9c665f-2be7-4b08-934d-b63036656b5f" containerID="219c63bcbbda1c8ee292c94dcc9e2f39b33c5aab0d972baaee0d1dc1f454692d" exitCode=143 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.197623 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-58f6cff88f-s86kk" event={"ID":"3d9c665f-2be7-4b08-934d-b63036656b5f","Type":"ContainerDied","Data":"219c63bcbbda1c8ee292c94dcc9e2f39b33c5aab0d972baaee0d1dc1f454692d"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.205238 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2c04aca3-abaa-4c62-8e62-af920276cc50/ovsdbserver-sb/0.log" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.205326 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2c04aca3-abaa-4c62-8e62-af920276cc50","Type":"ContainerDied","Data":"3cc7f342b1dff7530f9912e2834f34fe4b434b6e60ff2bc0acd8adfe099bd5bf"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.205436 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.217526 4944 generic.go:334] "Generic (PLEG): container finished" podID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerID="ceef1855e41bb1981a2f6e513be92fa9f3b4e90878c9354637a799ca0b3bbbf4" exitCode=143 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.217613 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce23facc-1a0d-4c53-b696-b37a7525eba7","Type":"ContainerDied","Data":"ceef1855e41bb1981a2f6e513be92fa9f3b4e90878c9354637a799ca0b3bbbf4"} Nov 24 09:14:24 crc kubenswrapper[4944]: E1124 09:14:24.218927 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.220459 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement0cdc-account-delete-sl7j9" event={"ID":"69996228-3cfa-486e-8291-12e44f438c5b","Type":"ContainerStarted","Data":"84d709e1a303c0866c88c4f8e95417f7ad9f12fe9bcb9c507763a9aae0a1ee4e"} Nov 24 09:14:24 crc kubenswrapper[4944]: E1124 09:14:24.223009 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 09:14:24 crc kubenswrapper[4944]: E1124 09:14:24.228519 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 09:14:24 crc kubenswrapper[4944]: E1124 09:14:24.228588 4944 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerName="ovn-northd" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.230998 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder575d-account-delete-8dqqk" event={"ID":"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2","Type":"ContainerStarted","Data":"5484dcdc1c11dd6f328aac9c91dc1ec5d3655a5649af2c5c9c971f820d6b01bc"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.238089 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance509f-account-delete-fx5bb"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.255795 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="709b6d2b-7e0f-4f60-9bd2-edacb409a727" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.193:6080/vnc_lite.html\": dial tcp 10.217.0.193:6080: connect: connection refused" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.259781 4944 generic.go:334] "Generic (PLEG): container finished" podID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerID="b626ffc5d738486bd96a728516fb7e89c1bf6800fb2f82fd0b1bc9e2f6d19268" exitCode=0 Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.260027 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" event={"ID":"42d1b549-198c-4a0d-a895-2b1f10dc337d","Type":"ContainerDied","Data":"b626ffc5d738486bd96a728516fb7e89c1bf6800fb2f82fd0b1bc9e2f6d19268"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.269599 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b9d45d4c-ab9f-426c-a193-7f46c398ed64/ovsdbserver-nb/0.log" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.269688 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b9d45d4c-ab9f-426c-a193-7f46c398ed64","Type":"ContainerDied","Data":"c38d9bc8e2ba464c90202d5ba1fe348490abf6a227c1249ebfab53d71639a12d"} Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.269780 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 09:14:24 crc kubenswrapper[4944]: E1124 09:14:24.293063 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 09:14:24 crc kubenswrapper[4944]: E1124 09:14:24.293144 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data podName:67e92d3f-3532-467f-bf38-c9c3107d4fcb nodeName:}" failed. No retries permitted until 2025-11-24 09:14:28.293125983 +0000 UTC m=+1328.827566455 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data") pod "rabbitmq-server-0" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb") : configmap "rabbitmq-config-data" not found Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.309823 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d3ec0c0-b512-4bca-aa21-57c13f34b5dd" path="/var/lib/kubelet/pods/3d3ec0c0-b512-4bca-aa21-57c13f34b5dd/volumes" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.310946 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ae35e54-8396-4a44-90d7-ae4afd3322af" path="/var/lib/kubelet/pods/5ae35e54-8396-4a44-90d7-ae4afd3322af/volumes" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.311888 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60b786e3-e5e4-4503-9887-9a3faef1d5c0" path="/var/lib/kubelet/pods/60b786e3-e5e4-4503-9887-9a3faef1d5c0/volumes" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.312870 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df085b2d-dbd0-4625-a905-a5f2c1dbedd8" path="/var/lib/kubelet/pods/df085b2d-dbd0-4625-a905-a5f2c1dbedd8/volumes" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.350487 4944 scope.go:117] "RemoveContainer" containerID="823389b569a8fd3fda96971a913dc73c72a4f283c5b7ab6145f8fb0dd9b6cb9e" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.399461 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.493885 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron7ac6-account-delete-7kth8"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.496679 4944 scope.go:117] "RemoveContainer" containerID="1eb23035ae5379ce721468ef5b81b6c60c5a92e85e392991180915600706763e" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.496968 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-combined-ca-bundle\") pod \"9a8dab89-ec08-4497-a890-bcb70c8ea430\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.497006 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config-secret\") pod \"9a8dab89-ec08-4497-a890-bcb70c8ea430\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.497040 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config\") pod \"9a8dab89-ec08-4497-a890-bcb70c8ea430\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.497940 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw9rf\" (UniqueName: \"kubernetes.io/projected/9a8dab89-ec08-4497-a890-bcb70c8ea430-kube-api-access-bw9rf\") pod \"9a8dab89-ec08-4497-a890-bcb70c8ea430\" (UID: \"9a8dab89-ec08-4497-a890-bcb70c8ea430\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.529534 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a8dab89-ec08-4497-a890-bcb70c8ea430-kube-api-access-bw9rf" (OuterVolumeSpecName: "kube-api-access-bw9rf") pod "9a8dab89-ec08-4497-a890-bcb70c8ea430" (UID: "9a8dab89-ec08-4497-a890-bcb70c8ea430"). InnerVolumeSpecName "kube-api-access-bw9rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.542175 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapie0a2-account-delete-bvsxd"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.543326 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "9a8dab89-ec08-4497-a890-bcb70c8ea430" (UID: "9a8dab89-ec08-4497-a890-bcb70c8ea430"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.608659 4944 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.608955 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw9rf\" (UniqueName: \"kubernetes.io/projected/9a8dab89-ec08-4497-a890-bcb70c8ea430-kube-api-access-bw9rf\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.617386 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rsh4b"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.628058 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-rsh4b"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.653576 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.672293 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.692142 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0eec8-account-delete-jtnrx"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.709129 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-npmfc"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.716704 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-npmfc"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.725156 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-m2b6x"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.730934 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a8dab89-ec08-4497-a890-bcb70c8ea430" (UID: "9a8dab89-ec08-4497-a890-bcb70c8ea430"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.741528 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.753656 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-m2b6x"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.767164 4944 scope.go:117] "RemoveContainer" containerID="b41e32588963f3e10325bdeeaf1f1bb9e18dd5fa9618690b49d86fb5f1d1c724" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.776032 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.797126 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.801629 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.804092 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "9a8dab89-ec08-4497-a890-bcb70c8ea430" (UID: "9a8dab89-ec08-4497-a890-bcb70c8ea430"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.812878 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.812914 4944 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9a8dab89-ec08-4497-a890-bcb70c8ea430-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:24 crc kubenswrapper[4944]: E1124 09:14:24.812991 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:24 crc kubenswrapper[4944]: E1124 09:14:24.813209 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data podName:12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5 nodeName:}" failed. No retries permitted until 2025-11-24 09:14:28.813030772 +0000 UTC m=+1329.347471234 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data") pod "rabbitmq-cell1-server-0" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5") : configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.893129 4944 scope.go:117] "RemoveContainer" containerID="878cbf2bbb403dd3a431aa6956e733ae81fc764310e89c3f03f0d40d6516dc09" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.913666 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-combined-ca-bundle\") pod \"e454a738-0872-41d8-9432-17432276248c\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.913830 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-operator-scripts\") pod \"e454a738-0872-41d8-9432-17432276248c\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.914003 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"e454a738-0872-41d8-9432-17432276248c\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.914105 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp7ql\" (UniqueName: \"kubernetes.io/projected/e454a738-0872-41d8-9432-17432276248c-kube-api-access-tp7ql\") pod \"e454a738-0872-41d8-9432-17432276248c\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.914240 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-nova-novncproxy-tls-certs\") pod \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.914415 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-config-data-default\") pod \"e454a738-0872-41d8-9432-17432276248c\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.914524 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-kolla-config\") pod \"e454a738-0872-41d8-9432-17432276248c\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.914615 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7mrb\" (UniqueName: \"kubernetes.io/projected/709b6d2b-7e0f-4f60-9bd2-edacb409a727-kube-api-access-g7mrb\") pod \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.914710 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-config-data\") pod \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.914814 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e454a738-0872-41d8-9432-17432276248c-config-data-generated\") pod \"e454a738-0872-41d8-9432-17432276248c\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.914965 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-galera-tls-certs\") pod \"e454a738-0872-41d8-9432-17432276248c\" (UID: \"e454a738-0872-41d8-9432-17432276248c\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.915100 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-vencrypt-tls-certs\") pod \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.915232 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-combined-ca-bundle\") pod \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\" (UID: \"709b6d2b-7e0f-4f60-9bd2-edacb409a727\") " Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.915944 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e454a738-0872-41d8-9432-17432276248c" (UID: "e454a738-0872-41d8-9432-17432276248c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.916372 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "e454a738-0872-41d8-9432-17432276248c" (UID: "e454a738-0872-41d8-9432-17432276248c"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.917592 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "e454a738-0872-41d8-9432-17432276248c" (UID: "e454a738-0872-41d8-9432-17432276248c"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.917879 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e454a738-0872-41d8-9432-17432276248c-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "e454a738-0872-41d8-9432-17432276248c" (UID: "e454a738-0872-41d8-9432-17432276248c"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.927923 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/709b6d2b-7e0f-4f60-9bd2-edacb409a727-kube-api-access-g7mrb" (OuterVolumeSpecName: "kube-api-access-g7mrb") pod "709b6d2b-7e0f-4f60-9bd2-edacb409a727" (UID: "709b6d2b-7e0f-4f60-9bd2-edacb409a727"). InnerVolumeSpecName "kube-api-access-g7mrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:24 crc kubenswrapper[4944]: I1124 09:14:24.937470 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e454a738-0872-41d8-9432-17432276248c-kube-api-access-tp7ql" (OuterVolumeSpecName: "kube-api-access-tp7ql") pod "e454a738-0872-41d8-9432-17432276248c" (UID: "e454a738-0872-41d8-9432-17432276248c"). InnerVolumeSpecName "kube-api-access-tp7ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.030618 4944 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.035573 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7mrb\" (UniqueName: \"kubernetes.io/projected/709b6d2b-7e0f-4f60-9bd2-edacb409a727-kube-api-access-g7mrb\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.035718 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e454a738-0872-41d8-9432-17432276248c-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.035798 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.035925 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp7ql\" (UniqueName: \"kubernetes.io/projected/e454a738-0872-41d8-9432-17432276248c-kube-api-access-tp7ql\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.036062 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e454a738-0872-41d8-9432-17432276248c-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.048086 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "e454a738-0872-41d8-9432-17432276248c" (UID: "e454a738-0872-41d8-9432-17432276248c"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.067194 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-config-data" (OuterVolumeSpecName: "config-data") pod "709b6d2b-7e0f-4f60-9bd2-edacb409a727" (UID: "709b6d2b-7e0f-4f60-9bd2-edacb409a727"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.111576 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e454a738-0872-41d8-9432-17432276248c" (UID: "e454a738-0872-41d8-9432-17432276248c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.139810 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.139865 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.139878 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.142295 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "709b6d2b-7e0f-4f60-9bd2-edacb409a727" (UID: "709b6d2b-7e0f-4f60-9bd2-edacb409a727"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.167096 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.171297 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "e454a738-0872-41d8-9432-17432276248c" (UID: "e454a738-0872-41d8-9432-17432276248c"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.232020 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "709b6d2b-7e0f-4f60-9bd2-edacb409a727" (UID: "709b6d2b-7e0f-4f60-9bd2-edacb409a727"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.242847 4944 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e454a738-0872-41d8-9432-17432276248c-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.243568 4944 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.243588 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.243600 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.274663 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "709b6d2b-7e0f-4f60-9bd2-edacb409a727" (UID: "709b6d2b-7e0f-4f60-9bd2-edacb409a727"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.284454 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"709b6d2b-7e0f-4f60-9bd2-edacb409a727","Type":"ContainerDied","Data":"1fa5d802adbb6046d2c8102e2c47bf4b2d3d25fb826929f651055f3fb472fb9e"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.284553 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.299376 4944 generic.go:334] "Generic (PLEG): container finished" podID="2b2d40c3-9522-46f4-85e7-fd724cb1bdf2" containerID="5e472b6966fd4a5f6d4239a02c06bce03ae0ec88a39cd86b701265cd0f16922a" exitCode=0 Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.299466 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder575d-account-delete-8dqqk" event={"ID":"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2","Type":"ContainerDied","Data":"5e472b6966fd4a5f6d4239a02c06bce03ae0ec88a39cd86b701265cd0f16922a"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.306639 4944 generic.go:334] "Generic (PLEG): container finished" podID="1db23d75-43f2-4bae-9e09-399d0ffb7ed5" containerID="41a16bdf6b3d1976b1e1aad714788e4ff1661da8136d3a26d52035543b68961c" exitCode=0 Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.306880 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance509f-account-delete-fx5bb" event={"ID":"1db23d75-43f2-4bae-9e09-399d0ffb7ed5","Type":"ContainerDied","Data":"41a16bdf6b3d1976b1e1aad714788e4ff1661da8136d3a26d52035543b68961c"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.306907 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance509f-account-delete-fx5bb" event={"ID":"1db23d75-43f2-4bae-9e09-399d0ffb7ed5","Type":"ContainerStarted","Data":"96fe8c2872672ff999266afb3d7e5a06e9872d7521b5424a28a5b52a0902127c"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.331732 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapie0a2-account-delete-bvsxd" event={"ID":"2a65aeee-1ac4-4ae3-b713-ad1d0074f797","Type":"ContainerStarted","Data":"be44c541c7e303f957c6328e9e73f5913708f7866af336580d068740794278c5"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.347792 4944 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/709b6d2b-7e0f-4f60-9bd2-edacb409a727-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.390909 4944 generic.go:334] "Generic (PLEG): container finished" podID="55a9098c-7257-46f7-888c-c4bb6f9421a5" containerID="f79b603137a9e552f12da861f283799163e55db755214ecf4ce6c275a2f775c7" exitCode=0 Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.391095 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicand270-account-delete-h2tcc" event={"ID":"55a9098c-7257-46f7-888c-c4bb6f9421a5","Type":"ContainerDied","Data":"f79b603137a9e552f12da861f283799163e55db755214ecf4ce6c275a2f775c7"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.410323 4944 generic.go:334] "Generic (PLEG): container finished" podID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerID="2c1a579178840a65cb60ad40f3b83a1c327413e6a2d77e8178e33176b65db789" exitCode=0 Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.410536 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" event={"ID":"42d1b549-198c-4a0d-a895-2b1f10dc337d","Type":"ContainerDied","Data":"2c1a579178840a65cb60ad40f3b83a1c327413e6a2d77e8178e33176b65db789"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.410577 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" event={"ID":"42d1b549-198c-4a0d-a895-2b1f10dc337d","Type":"ContainerDied","Data":"94bb71dd1ae49e891af53f55df27e1e816740e7be55a0f56d91419eaf28f8148"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.410589 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94bb71dd1ae49e891af53f55df27e1e816740e7be55a0f56d91419eaf28f8148" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.421821 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0eec8-account-delete-jtnrx" event={"ID":"b69c572d-53a7-4d39-a8fc-4152846915a5","Type":"ContainerStarted","Data":"11f5eb7d3d7c556ad26a6226db3ba58bc6b217dd6a967125a016d0a426fa34fd"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.427315 4944 generic.go:334] "Generic (PLEG): container finished" podID="e454a738-0872-41d8-9432-17432276248c" containerID="06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d" exitCode=0 Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.427401 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e454a738-0872-41d8-9432-17432276248c","Type":"ContainerDied","Data":"06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.427434 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e454a738-0872-41d8-9432-17432276248c","Type":"ContainerDied","Data":"6236aed52bc908141599bf898d07bc26522c296edb4567ff1515a8486312f72f"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.427534 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.432489 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement0cdc-account-delete-sl7j9" event={"ID":"69996228-3cfa-486e-8291-12e44f438c5b","Type":"ContainerStarted","Data":"baeed649469760ea552380de706d9f29c4df36ac76ccb2186b7193cf50ccbe23"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.437708 4944 generic.go:334] "Generic (PLEG): container finished" podID="98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32" containerID="83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b" exitCode=0 Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.437794 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32","Type":"ContainerDied","Data":"83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.439605 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.449875 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron7ac6-account-delete-7kth8" event={"ID":"11d8c8f9-ee83-4450-9a76-3c0a020652de","Type":"ContainerStarted","Data":"8de6f9b2902383bc519a27bcb1dbae6df05a0319181cea964b80d672e04e2480"} Nov 24 09:14:25 crc kubenswrapper[4944]: I1124 09:14:25.499984 4944 scope.go:117] "RemoveContainer" containerID="61ee3036c3b98fca180967e52f25b5128d65397ebffba1911bb8a5e25c2ec958" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.545084 4944 scope.go:117] "RemoveContainer" containerID="3ce79850e6073ff3d7e81da5fd81b8be25dfceb1d21948518006ccbe0b1eda7a" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.623939 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:14:26 crc kubenswrapper[4944]: E1124 09:14:25.687142 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b is running failed: container process not found" containerID="83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 09:14:26 crc kubenswrapper[4944]: E1124 09:14:25.688753 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b is running failed: container process not found" containerID="83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 09:14:26 crc kubenswrapper[4944]: E1124 09:14:25.689144 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b is running failed: container process not found" containerID="83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 09:14:26 crc kubenswrapper[4944]: E1124 09:14:25.689181 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32" containerName="nova-cell1-conductor-conductor" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.697086 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.708330 4944 scope.go:117] "RemoveContainer" containerID="5ef21e16c9d939c2affb5458115459c6969a680e9f313ed1d7b13c1de4f3f96a" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.729841 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.747439 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.756660 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.761240 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-combined-ca-bundle\") pod \"42d1b549-198c-4a0d-a895-2b1f10dc337d\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.761295 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-run-httpd\") pod \"42d1b549-198c-4a0d-a895-2b1f10dc337d\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.761374 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qtfg\" (UniqueName: \"kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-kube-api-access-4qtfg\") pod \"42d1b549-198c-4a0d-a895-2b1f10dc337d\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.761471 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-public-tls-certs\") pod \"42d1b549-198c-4a0d-a895-2b1f10dc337d\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.761560 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-etc-swift\") pod \"42d1b549-198c-4a0d-a895-2b1f10dc337d\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.761589 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-log-httpd\") pod \"42d1b549-198c-4a0d-a895-2b1f10dc337d\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.761630 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-config-data\") pod \"42d1b549-198c-4a0d-a895-2b1f10dc337d\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.761675 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-internal-tls-certs\") pod \"42d1b549-198c-4a0d-a895-2b1f10dc337d\" (UID: \"42d1b549-198c-4a0d-a895-2b1f10dc337d\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.763221 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "42d1b549-198c-4a0d-a895-2b1f10dc337d" (UID: "42d1b549-198c-4a0d-a895-2b1f10dc337d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.764990 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "42d1b549-198c-4a0d-a895-2b1f10dc337d" (UID: "42d1b549-198c-4a0d-a895-2b1f10dc337d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.772673 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.784595 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "42d1b549-198c-4a0d-a895-2b1f10dc337d" (UID: "42d1b549-198c-4a0d-a895-2b1f10dc337d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.796985 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicand270-account-delete-h2tcc" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.818972 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-kube-api-access-4qtfg" (OuterVolumeSpecName: "kube-api-access-4qtfg") pod "42d1b549-198c-4a0d-a895-2b1f10dc337d" (UID: "42d1b549-198c-4a0d-a895-2b1f10dc337d"). InnerVolumeSpecName "kube-api-access-4qtfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.863963 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-combined-ca-bundle\") pod \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.864235 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-config-data\") pod \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.864421 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hdvk\" (UniqueName: \"kubernetes.io/projected/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-kube-api-access-5hdvk\") pod \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\" (UID: \"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.865296 4944 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.865314 4944 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.865323 4944 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d1b549-198c-4a0d-a895-2b1f10dc337d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.865333 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qtfg\" (UniqueName: \"kubernetes.io/projected/42d1b549-198c-4a0d-a895-2b1f10dc337d-kube-api-access-4qtfg\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.868456 4944 scope.go:117] "RemoveContainer" containerID="bc869a248132ca4279a0e2f68eec18b02bc447d39b21ce0e334fb974d4d68487" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.883165 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-kube-api-access-5hdvk" (OuterVolumeSpecName: "kube-api-access-5hdvk") pod "98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32" (UID: "98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32"). InnerVolumeSpecName "kube-api-access-5hdvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.920669 4944 scope.go:117] "RemoveContainer" containerID="06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d" Nov 24 09:14:26 crc kubenswrapper[4944]: E1124 09:14:25.931916 4944 kuberuntime_gc.go:389] "Failed to remove container log dead symlink" err="remove /var/log/containers/openstack-cell1-galera-0_openstack_galera-06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d.log: no such file or directory" path="/var/log/containers/openstack-cell1-galera-0_openstack_galera-06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d.log" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.968629 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a9098c-7257-46f7-888c-c4bb6f9421a5-operator-scripts\") pod \"55a9098c-7257-46f7-888c-c4bb6f9421a5\" (UID: \"55a9098c-7257-46f7-888c-c4bb6f9421a5\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.968734 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xq5x6\" (UniqueName: \"kubernetes.io/projected/55a9098c-7257-46f7-888c-c4bb6f9421a5-kube-api-access-xq5x6\") pod \"55a9098c-7257-46f7-888c-c4bb6f9421a5\" (UID: \"55a9098c-7257-46f7-888c-c4bb6f9421a5\") " Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.969778 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55a9098c-7257-46f7-888c-c4bb6f9421a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "55a9098c-7257-46f7-888c-c4bb6f9421a5" (UID: "55a9098c-7257-46f7-888c-c4bb6f9421a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.969962 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a9098c-7257-46f7-888c-c4bb6f9421a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.969975 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hdvk\" (UniqueName: \"kubernetes.io/projected/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-kube-api-access-5hdvk\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:25.976858 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55a9098c-7257-46f7-888c-c4bb6f9421a5-kube-api-access-xq5x6" (OuterVolumeSpecName: "kube-api-access-xq5x6") pod "55a9098c-7257-46f7-888c-c4bb6f9421a5" (UID: "55a9098c-7257-46f7-888c-c4bb6f9421a5"). InnerVolumeSpecName "kube-api-access-xq5x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.020130 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32" (UID: "98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.040485 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-config-data" (OuterVolumeSpecName: "config-data") pod "42d1b549-198c-4a0d-a895-2b1f10dc337d" (UID: "42d1b549-198c-4a0d-a895-2b1f10dc337d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.075731 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xq5x6\" (UniqueName: \"kubernetes.io/projected/55a9098c-7257-46f7-888c-c4bb6f9421a5-kube-api-access-xq5x6\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.075765 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.075779 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.082472 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "42d1b549-198c-4a0d-a895-2b1f10dc337d" (UID: "42d1b549-198c-4a0d-a895-2b1f10dc337d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.124740 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-config-data" (OuterVolumeSpecName: "config-data") pod "98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32" (UID: "98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.173059 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.173355 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="ceilometer-central-agent" containerID="cri-o://73095b06c9c5460d64ea3a6e6ad293c16acb97dedadc392b482d5b45673686f5" gracePeriod=30 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.173743 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="proxy-httpd" containerID="cri-o://a84b239223662f1225b37375264550577380c4e7d01f7c269035f7022dd18d82" gracePeriod=30 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.173789 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="sg-core" containerID="cri-o://dbee395d02885c9a06d22bd217f37d54477844b3e398fa6029fcb08eebb7be4e" gracePeriod=30 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.173822 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="ceilometer-notification-agent" containerID="cri-o://f1c621c73eb738098cb49670f9fd9711e1d1a037fa44d09a782be7678f0546b3" gracePeriod=30 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.177537 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.177571 4944 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.235516 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "42d1b549-198c-4a0d-a895-2b1f10dc337d" (UID: "42d1b549-198c-4a0d-a895-2b1f10dc337d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.252110 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.252316 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="d2e0b8db-ac91-432d-a87c-e741e441c33a" containerName="kube-state-metrics" containerID="cri-o://cb293c82a7b8cd484e5348b25efb88a1b9008ec1eeee27738a2aeda8a4a18930" gracePeriod=30 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.279882 4944 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.286708 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:55110->10.217.0.200:8775: read: connection reset by peer" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.292253 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:55108->10.217.0.200:8775: read: connection reset by peer" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.325288 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42d1b549-198c-4a0d-a895-2b1f10dc337d" (UID: "42d1b549-198c-4a0d-a895-2b1f10dc337d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.335478 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c04aca3-abaa-4c62-8e62-af920276cc50" path="/var/lib/kubelet/pods/2c04aca3-abaa-4c62-8e62-af920276cc50/volumes" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.336298 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" path="/var/lib/kubelet/pods/5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef/volumes" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.336918 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69084f7b-b347-4fdf-917e-2f534f3cc47c" path="/var/lib/kubelet/pods/69084f7b-b347-4fdf-917e-2f534f3cc47c/volumes" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.338128 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="709b6d2b-7e0f-4f60-9bd2-edacb409a727" path="/var/lib/kubelet/pods/709b6d2b-7e0f-4f60-9bd2-edacb409a727/volumes" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.338687 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a8dab89-ec08-4497-a890-bcb70c8ea430" path="/var/lib/kubelet/pods/9a8dab89-ec08-4497-a890-bcb70c8ea430/volumes" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.339520 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9d45d4c-ab9f-426c-a193-7f46c398ed64" path="/var/lib/kubelet/pods/b9d45d4c-ab9f-426c-a193-7f46c398ed64/volumes" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.344368 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c521e9a3-0d64-4658-8a00-690d20e619d0" path="/var/lib/kubelet/pods/c521e9a3-0d64-4658-8a00-690d20e619d0/volumes" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.345234 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e454a738-0872-41d8-9432-17432276248c" path="/var/lib/kubelet/pods/e454a738-0872-41d8-9432-17432276248c/volumes" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.374929 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.375161 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="35137b92-4f56-4399-8121-89bdb126162b" containerName="memcached" containerID="cri-o://10473e2e58eff197e7d6054c78b1b43c310d54ae418a52fb6168c9c0b59133e8" gracePeriod=30 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.396593 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d1b549-198c-4a0d-a895-2b1f10dc337d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.454123 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-79qbw"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.469742 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-79qbw"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.486668 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-75jn8"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.495469 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-75jn8"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.508198 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5457b9857f-kg9gm"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.508399 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-5457b9857f-kg9gm" podUID="d2495f8a-6f01-47e6-beff-f0d7345967a7" containerName="keystone-api" containerID="cri-o://5950163d1a4608641fb113f50f6737d5436fc15138c0f58c9d9ec29a06eda578" gracePeriod=30 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.509317 4944 generic.go:334] "Generic (PLEG): container finished" podID="b9a8d972-b16f-421e-9fe4-f57653048845" containerID="84fadfabdc280facbbd9a3442d721a8b2b1c0298a2c5bd828ac25612a83c8d81" exitCode=0 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.509362 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a8d972-b16f-421e-9fe4-f57653048845","Type":"ContainerDied","Data":"84fadfabdc280facbbd9a3442d721a8b2b1c0298a2c5bd828ac25612a83c8d81"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.538300 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.549749 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="b9a8d972-b16f-421e-9fe4-f57653048845" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.166:8776/healthcheck\": dial tcp 10.217.0.166:8776: connect: connection refused" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.559695 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-cw4g5"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.569977 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-cw4g5"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.576597 4944 generic.go:334] "Generic (PLEG): container finished" podID="70a59d46-95bf-4681-b61d-22239c638737" containerID="e42a48dc5f9a748005eb91765bdfa38b2ad4b4f974b03b15a72a6b00023e47a2" exitCode=0 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.576653 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5ccb7b4f9b-mcdmd" event={"ID":"70a59d46-95bf-4681-b61d-22239c638737","Type":"ContainerDied","Data":"e42a48dc5f9a748005eb91765bdfa38b2ad4b4f974b03b15a72a6b00023e47a2"} Nov 24 09:14:26 crc kubenswrapper[4944]: E1124 09:14:26.584585 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249 is running failed: container process not found" containerID="58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.592567 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0eec8-account-delete-jtnrx"] Nov 24 09:14:26 crc kubenswrapper[4944]: E1124 09:14:26.611333 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249 is running failed: container process not found" containerID="58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.612182 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-eec8-account-create-snmtp"] Nov 24 09:14:26 crc kubenswrapper[4944]: E1124 09:14:26.616930 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249 is running failed: container process not found" containerID="58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 09:14:26 crc kubenswrapper[4944]: E1124 09:14:26.616987 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="90fb3cc2-0725-48dc-9ad5-5c8da27930f8" containerName="nova-scheduler-scheduler" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.631941 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-eec8-account-create-snmtp"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.637180 4944 generic.go:334] "Generic (PLEG): container finished" podID="7209b8a6-5a97-4611-8183-84f1db25e2d1" containerID="71ca2485422737d39ab2f15fc0ad24cd930f9cf353b7e908277d85b6a71d3f14" exitCode=0 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.637317 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7209b8a6-5a97-4611-8183-84f1db25e2d1","Type":"ContainerDied","Data":"71ca2485422737d39ab2f15fc0ad24cd930f9cf353b7e908277d85b6a71d3f14"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.659715 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-wgxhq"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.667367 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-wgxhq"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.670479 4944 generic.go:334] "Generic (PLEG): container finished" podID="90fb3cc2-0725-48dc-9ad5-5c8da27930f8" containerID="58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249" exitCode=0 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.671235 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"90fb3cc2-0725-48dc-9ad5-5c8da27930f8","Type":"ContainerDied","Data":"58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.677551 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-1ab7-account-create-c4ljx"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.677696 4944 generic.go:334] "Generic (PLEG): container finished" podID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerID="dbee395d02885c9a06d22bd217f37d54477844b3e398fa6029fcb08eebb7be4e" exitCode=2 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.677761 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4","Type":"ContainerDied","Data":"dbee395d02885c9a06d22bd217f37d54477844b3e398fa6029fcb08eebb7be4e"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.684489 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-1ab7-account-create-c4ljx"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.687060 4944 generic.go:334] "Generic (PLEG): container finished" podID="a8d71630-681e-4126-bb64-e8b03f83ce97" containerID="2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef" exitCode=0 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.687217 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8d71630-681e-4126-bb64-e8b03f83ce97","Type":"ContainerDied","Data":"2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.692856 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.693262 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32","Type":"ContainerDied","Data":"042d6d603e1db607911a0cb7bc01d5667ef75d0890485349d0ab3b3db6a46948"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.695073 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicand270-account-delete-h2tcc" event={"ID":"55a9098c-7257-46f7-888c-c4bb6f9421a5","Type":"ContainerDied","Data":"18abcd5217b86e8d889c6cc23c83d4771c98be7cd169a305a78158c2bd5ad07b"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.695094 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18abcd5217b86e8d889c6cc23c83d4771c98be7cd169a305a78158c2bd5ad07b" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.695161 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicand270-account-delete-h2tcc" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.699905 4944 generic.go:334] "Generic (PLEG): container finished" podID="2a65aeee-1ac4-4ae3-b713-ad1d0074f797" containerID="87bc89b8a47e19382a1d4524937f8b6a5ab31d4623f694f16b7b41a3fda03f50" exitCode=0 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.699968 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapie0a2-account-delete-bvsxd" event={"ID":"2a65aeee-1ac4-4ae3-b713-ad1d0074f797","Type":"ContainerDied","Data":"87bc89b8a47e19382a1d4524937f8b6a5ab31d4623f694f16b7b41a3fda03f50"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.702132 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0eec8-account-delete-jtnrx" event={"ID":"b69c572d-53a7-4d39-a8fc-4152846915a5","Type":"ContainerStarted","Data":"0e567cd20074605f8fc1922dc6cc36a075eb716e38a58a3ff08e0c19da6b0f70"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.713933 4944 generic.go:334] "Generic (PLEG): container finished" podID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerID="f4fda6a5047a5412abc68148082d7806e36494cd3997be68245586f663ec9996" exitCode=0 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.713982 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc12ffc5-e026-42a6-8ddc-43961c535787","Type":"ContainerDied","Data":"f4fda6a5047a5412abc68148082d7806e36494cd3997be68245586f663ec9996"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.718336 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron7ac6-account-delete-7kth8" event={"ID":"11d8c8f9-ee83-4450-9a76-3c0a020652de","Type":"ContainerStarted","Data":"873fb4ad2a63dd371217f9b304da3a326d22fb85587ad6f5c65d2bd833dca7af"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.740489 4944 generic.go:334] "Generic (PLEG): container finished" podID="69996228-3cfa-486e-8291-12e44f438c5b" containerID="baeed649469760ea552380de706d9f29c4df36ac76ccb2186b7193cf50ccbe23" exitCode=0 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.740574 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement0cdc-account-delete-sl7j9" event={"ID":"69996228-3cfa-486e-8291-12e44f438c5b","Type":"ContainerDied","Data":"baeed649469760ea552380de706d9f29c4df36ac76ccb2186b7193cf50ccbe23"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.774931 4944 generic.go:334] "Generic (PLEG): container finished" podID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerID="1156a54f88c1c2dd0577bc2b082f7256b862e090c82e8172fc3df6b5bf9e5bdf" exitCode=0 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.775700 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce23facc-1a0d-4c53-b696-b37a7525eba7","Type":"ContainerDied","Data":"1156a54f88c1c2dd0577bc2b082f7256b862e090c82e8172fc3df6b5bf9e5bdf"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.782365 4944 generic.go:334] "Generic (PLEG): container finished" podID="f94b9960-fa37-49f7-9fa7-0e01f855992b" containerID="9a5211b2a8ad4ed0c93e5f29ec422b4395da812067c859ba887bf13db9c4afb6" exitCode=0 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.782423 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f94b9960-fa37-49f7-9fa7-0e01f855992b","Type":"ContainerDied","Data":"9a5211b2a8ad4ed0c93e5f29ec422b4395da812067c859ba887bf13db9c4afb6"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.788459 4944 generic.go:334] "Generic (PLEG): container finished" podID="d2e0b8db-ac91-432d-a87c-e741e441c33a" containerID="cb293c82a7b8cd484e5348b25efb88a1b9008ec1eeee27738a2aeda8a4a18930" exitCode=2 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.788573 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d2e0b8db-ac91-432d-a87c-e741e441c33a","Type":"ContainerDied","Data":"cb293c82a7b8cd484e5348b25efb88a1b9008ec1eeee27738a2aeda8a4a18930"} Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.788674 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.801840 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.812235 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.815662 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="18d26ea6-eff1-4ad5-b50b-dd233a22e982" containerName="galera" containerID="cri-o://1a95527c06d2048bd5100167ef2d1b762657135708fcae5c22fd4931034b3057" gracePeriod=30 Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.822591 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7bcd854fd7-z2d6k"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.857636 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-7bcd854fd7-z2d6k"] Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.872031 4944 scope.go:117] "RemoveContainer" containerID="f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.907456 4944 scope.go:117] "RemoveContainer" containerID="06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d" Nov 24 09:14:26 crc kubenswrapper[4944]: E1124 09:14:26.908165 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d\": container with ID starting with 06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d not found: ID does not exist" containerID="06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.908210 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d"} err="failed to get container status \"06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d\": rpc error: code = NotFound desc = could not find container \"06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d\": container with ID starting with 06522627ceaa5439e3f74fdd34f1d8050bc71a145b266f2a8550a4f2de41200d not found: ID does not exist" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.908241 4944 scope.go:117] "RemoveContainer" containerID="f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c" Nov 24 09:14:26 crc kubenswrapper[4944]: E1124 09:14:26.909022 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c\": container with ID starting with f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c not found: ID does not exist" containerID="f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.909085 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c"} err="failed to get container status \"f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c\": rpc error: code = NotFound desc = could not find container \"f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c\": container with ID starting with f6417930c2379ed534d690117d84bd8da6a63440dfcd15ae64080273618f3c1c not found: ID does not exist" Nov 24 09:14:26 crc kubenswrapper[4944]: I1124 09:14:26.909119 4944 scope.go:117] "RemoveContainer" containerID="83c3204692cef396ec88f75a06f819f6f6d3284f6ca821ae380c13dca5594a6b" Nov 24 09:14:27 crc kubenswrapper[4944]: I1124 09:14:27.045470 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": dial tcp 10.217.0.195:3000: connect: connection refused" Nov 24 09:14:27 crc kubenswrapper[4944]: I1124 09:14:27.267692 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cd5cbd7b9-rsh4b" podUID="5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.194:5353: i/o timeout" Nov 24 09:14:27 crc kubenswrapper[4944]: E1124 09:14:27.897861 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:27 crc kubenswrapper[4944]: I1124 09:14:27.899643 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="35137b92-4f56-4399-8121-89bdb126162b" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.105:11211: connect: connection refused" Nov 24 09:14:27 crc kubenswrapper[4944]: E1124 09:14:27.904644 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef is running failed: container process not found" containerID="2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 09:14:27 crc kubenswrapper[4944]: E1124 09:14:27.904757 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:27 crc kubenswrapper[4944]: E1124 09:14:27.911716 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef is running failed: container process not found" containerID="2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 09:14:27 crc kubenswrapper[4944]: E1124 09:14:27.911862 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:27 crc kubenswrapper[4944]: E1124 09:14:27.911887 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server" Nov 24 09:14:27 crc kubenswrapper[4944]: I1124 09:14:27.913795 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f49db88bb-dzrb6" podUID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.161:9311/healthcheck\": dial tcp 10.217.0.161:9311: connect: connection refused" Nov 24 09:14:27 crc kubenswrapper[4944]: I1124 09:14:27.914154 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f49db88bb-dzrb6" podUID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.161:9311/healthcheck\": dial tcp 10.217.0.161:9311: connect: connection refused" Nov 24 09:14:28 crc kubenswrapper[4944]: E1124 09:14:27.921210 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef is running failed: container process not found" containerID="2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 09:14:28 crc kubenswrapper[4944]: E1124 09:14:27.921343 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="a8d71630-681e-4126-bb64-e8b03f83ce97" containerName="nova-cell0-conductor-conductor" Nov 24 09:14:28 crc kubenswrapper[4944]: E1124 09:14:27.925856 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:28 crc kubenswrapper[4944]: E1124 09:14:27.934226 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:27.955403 4944 generic.go:334] "Generic (PLEG): container finished" podID="b69c572d-53a7-4d39-a8fc-4152846915a5" containerID="0e567cd20074605f8fc1922dc6cc36a075eb716e38a58a3ff08e0c19da6b0f70" exitCode=0 Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:27.955465 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0eec8-account-delete-jtnrx" event={"ID":"b69c572d-53a7-4d39-a8fc-4152846915a5","Type":"ContainerDied","Data":"0e567cd20074605f8fc1922dc6cc36a075eb716e38a58a3ff08e0c19da6b0f70"} Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:27.962764 4944 generic.go:334] "Generic (PLEG): container finished" podID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerID="7159a46717426d269e3e9ca766c99432598e2da70a9cf82e6a43abda53ec27b9" exitCode=0 Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:27.962816 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f49db88bb-dzrb6" event={"ID":"046a8c15-6acd-47fe-aef7-daac5085c2ec","Type":"ContainerDied","Data":"7159a46717426d269e3e9ca766c99432598e2da70a9cf82e6a43abda53ec27b9"} Nov 24 09:14:28 crc kubenswrapper[4944]: E1124 09:14:28.015749 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:28 crc kubenswrapper[4944]: E1124 09:14:28.015813 4944 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovs-vswitchd" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.110515 4944 generic.go:334] "Generic (PLEG): container finished" podID="11d8c8f9-ee83-4450-9a76-3c0a020652de" containerID="873fb4ad2a63dd371217f9b304da3a326d22fb85587ad6f5c65d2bd833dca7af" exitCode=0 Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.110608 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron7ac6-account-delete-7kth8" event={"ID":"11d8c8f9-ee83-4450-9a76-3c0a020652de","Type":"ContainerDied","Data":"873fb4ad2a63dd371217f9b304da3a326d22fb85587ad6f5c65d2bd833dca7af"} Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.190364 4944 generic.go:334] "Generic (PLEG): container finished" podID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerID="a84b239223662f1225b37375264550577380c4e7d01f7c269035f7022dd18d82" exitCode=0 Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.190405 4944 generic.go:334] "Generic (PLEG): container finished" podID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerID="73095b06c9c5460d64ea3a6e6ad293c16acb97dedadc392b482d5b45673686f5" exitCode=0 Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.190628 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4","Type":"ContainerDied","Data":"a84b239223662f1225b37375264550577380c4e7d01f7c269035f7022dd18d82"} Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.190662 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4","Type":"ContainerDied","Data":"73095b06c9c5460d64ea3a6e6ad293c16acb97dedadc392b482d5b45673686f5"} Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.310378 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19055ab7-c910-494d-9a31-49aadf127c69" path="/var/lib/kubelet/pods/19055ab7-c910-494d-9a31-49aadf127c69/volumes" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.311290 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cc3188b-7842-4aa0-accd-cfa2e303f001" path="/var/lib/kubelet/pods/1cc3188b-7842-4aa0-accd-cfa2e303f001/volumes" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.311773 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d229255-c311-4626-be89-4769eca20010" path="/var/lib/kubelet/pods/1d229255-c311-4626-be89-4769eca20010/volumes" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.313163 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42d1b549-198c-4a0d-a895-2b1f10dc337d" path="/var/lib/kubelet/pods/42d1b549-198c-4a0d-a895-2b1f10dc337d/volumes" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.316421 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32" path="/var/lib/kubelet/pods/98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32/volumes" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.317058 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b92b104c-e53e-4246-b5d0-099005284506" path="/var/lib/kubelet/pods/b92b104c-e53e-4246-b5d0-099005284506/volumes" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.317552 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b935dbba-ca03-419e-95de-f1061fbc0999" path="/var/lib/kubelet/pods/b935dbba-ca03-419e-95de-f1061fbc0999/volumes" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.323210 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d07bba62-af8e-415e-a68d-ed9bf7602bd6" path="/var/lib/kubelet/pods/d07bba62-af8e-415e-a68d-ed9bf7602bd6/volumes" Nov 24 09:14:28 crc kubenswrapper[4944]: E1124 09:14:28.326542 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 24 09:14:28 crc kubenswrapper[4944]: E1124 09:14:28.326618 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data podName:67e92d3f-3532-467f-bf38-c9c3107d4fcb nodeName:}" failed. No retries permitted until 2025-11-24 09:14:36.326599216 +0000 UTC m=+1336.861039678 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data") pod "rabbitmq-server-0" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb") : configmap "rabbitmq-config-data" not found Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.335803 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.355035 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.368982 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.383332 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.384306 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.413578 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427300 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-combined-ca-bundle\") pod \"70a59d46-95bf-4681-b61d-22239c638737\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427370 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-config\") pod \"d2e0b8db-ac91-432d-a87c-e741e441c33a\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427408 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-combined-ca-bundle\") pod \"d2e0b8db-ac91-432d-a87c-e741e441c33a\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427449 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-internal-tls-certs\") pod \"70a59d46-95bf-4681-b61d-22239c638737\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427468 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70a59d46-95bf-4681-b61d-22239c638737-logs\") pod \"70a59d46-95bf-4681-b61d-22239c638737\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427526 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-combined-ca-bundle\") pod \"a8d71630-681e-4126-bb64-e8b03f83ce97\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427567 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-certs\") pod \"d2e0b8db-ac91-432d-a87c-e741e441c33a\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427603 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-config-data\") pod \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427639 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67gww\" (UniqueName: \"kubernetes.io/projected/a8d71630-681e-4126-bb64-e8b03f83ce97-kube-api-access-67gww\") pod \"a8d71630-681e-4126-bb64-e8b03f83ce97\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427686 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-combined-ca-bundle\") pod \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427714 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-public-tls-certs\") pod \"70a59d46-95bf-4681-b61d-22239c638737\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427776 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqqx7\" (UniqueName: \"kubernetes.io/projected/fc12ffc5-e026-42a6-8ddc-43961c535787-kube-api-access-zqqx7\") pod \"fc12ffc5-e026-42a6-8ddc-43961c535787\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427795 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-combined-ca-bundle\") pod \"fc12ffc5-e026-42a6-8ddc-43961c535787\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427827 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-scripts\") pod \"70a59d46-95bf-4681-b61d-22239c638737\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427844 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl4bh\" (UniqueName: \"kubernetes.io/projected/70a59d46-95bf-4681-b61d-22239c638737-kube-api-access-gl4bh\") pod \"70a59d46-95bf-4681-b61d-22239c638737\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427883 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-config-data\") pod \"70a59d46-95bf-4681-b61d-22239c638737\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427927 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-nova-metadata-tls-certs\") pod \"fc12ffc5-e026-42a6-8ddc-43961c535787\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.427983 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-config-data\") pod \"fc12ffc5-e026-42a6-8ddc-43961c535787\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.428024 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvbj6\" (UniqueName: \"kubernetes.io/projected/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-kube-api-access-hvbj6\") pod \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\" (UID: \"90fb3cc2-0725-48dc-9ad5-5c8da27930f8\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.428068 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf68g\" (UniqueName: \"kubernetes.io/projected/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-api-access-qf68g\") pod \"d2e0b8db-ac91-432d-a87c-e741e441c33a\" (UID: \"d2e0b8db-ac91-432d-a87c-e741e441c33a\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.428094 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-config-data\") pod \"a8d71630-681e-4126-bb64-e8b03f83ce97\" (UID: \"a8d71630-681e-4126-bb64-e8b03f83ce97\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.428120 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc12ffc5-e026-42a6-8ddc-43961c535787-logs\") pod \"fc12ffc5-e026-42a6-8ddc-43961c535787\" (UID: \"fc12ffc5-e026-42a6-8ddc-43961c535787\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.430810 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc12ffc5-e026-42a6-8ddc-43961c535787-logs" (OuterVolumeSpecName: "logs") pod "fc12ffc5-e026-42a6-8ddc-43961c535787" (UID: "fc12ffc5-e026-42a6-8ddc-43961c535787"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.447255 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70a59d46-95bf-4681-b61d-22239c638737-logs" (OuterVolumeSpecName: "logs") pod "70a59d46-95bf-4681-b61d-22239c638737" (UID: "70a59d46-95bf-4681-b61d-22239c638737"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.452025 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.452794 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-api-access-qf68g" (OuterVolumeSpecName: "kube-api-access-qf68g") pod "d2e0b8db-ac91-432d-a87c-e741e441c33a" (UID: "d2e0b8db-ac91-432d-a87c-e741e441c33a"). InnerVolumeSpecName "kube-api-access-qf68g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.455568 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.470538 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.526924 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-scripts" (OuterVolumeSpecName: "scripts") pod "70a59d46-95bf-4681-b61d-22239c638737" (UID: "70a59d46-95bf-4681-b61d-22239c638737"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.526945 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-kube-api-access-hvbj6" (OuterVolumeSpecName: "kube-api-access-hvbj6") pod "90fb3cc2-0725-48dc-9ad5-5c8da27930f8" (UID: "90fb3cc2-0725-48dc-9ad5-5c8da27930f8"). InnerVolumeSpecName "kube-api-access-hvbj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.529863 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-public-tls-certs\") pod \"ce23facc-1a0d-4c53-b696-b37a7525eba7\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.529942 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data\") pod \"b9a8d972-b16f-421e-9fe4-f57653048845\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.529969 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-config-data\") pod \"ce23facc-1a0d-4c53-b696-b37a7525eba7\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.530010 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-internal-tls-certs\") pod \"7209b8a6-5a97-4611-8183-84f1db25e2d1\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.530035 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"f94b9960-fa37-49f7-9fa7-0e01f855992b\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.530126 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p56z\" (UniqueName: \"kubernetes.io/projected/f94b9960-fa37-49f7-9fa7-0e01f855992b-kube-api-access-9p56z\") pod \"f94b9960-fa37-49f7-9fa7-0e01f855992b\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.530149 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-scripts\") pod \"f94b9960-fa37-49f7-9fa7-0e01f855992b\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.530725 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data-custom\") pod \"b9a8d972-b16f-421e-9fe4-f57653048845\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.530756 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-public-tls-certs\") pod \"f94b9960-fa37-49f7-9fa7-0e01f855992b\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.530805 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-scripts\") pod \"b9a8d972-b16f-421e-9fe4-f57653048845\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.530832 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-logs\") pod \"f94b9960-fa37-49f7-9fa7-0e01f855992b\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.530903 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-combined-ca-bundle\") pod \"7209b8a6-5a97-4611-8183-84f1db25e2d1\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.530936 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-config-data\") pod \"7209b8a6-5a97-4611-8183-84f1db25e2d1\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.530970 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-logs\") pod \"7209b8a6-5a97-4611-8183-84f1db25e2d1\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531026 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-internal-tls-certs\") pod \"b9a8d972-b16f-421e-9fe4-f57653048845\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531067 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-combined-ca-bundle\") pod \"ce23facc-1a0d-4c53-b696-b37a7525eba7\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531090 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a8d972-b16f-421e-9fe4-f57653048845-logs\") pod \"b9a8d972-b16f-421e-9fe4-f57653048845\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531114 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk7nh\" (UniqueName: \"kubernetes.io/projected/b9a8d972-b16f-421e-9fe4-f57653048845-kube-api-access-xk7nh\") pod \"b9a8d972-b16f-421e-9fe4-f57653048845\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531138 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"7209b8a6-5a97-4611-8183-84f1db25e2d1\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531176 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9kp5\" (UniqueName: \"kubernetes.io/projected/ce23facc-1a0d-4c53-b696-b37a7525eba7-kube-api-access-z9kp5\") pod \"ce23facc-1a0d-4c53-b696-b37a7525eba7\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531204 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-scripts\") pod \"7209b8a6-5a97-4611-8183-84f1db25e2d1\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531230 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-internal-tls-certs\") pod \"ce23facc-1a0d-4c53-b696-b37a7525eba7\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531256 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-combined-ca-bundle\") pod \"f94b9960-fa37-49f7-9fa7-0e01f855992b\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531278 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-httpd-run\") pod \"f94b9960-fa37-49f7-9fa7-0e01f855992b\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531297 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-combined-ca-bundle\") pod \"b9a8d972-b16f-421e-9fe4-f57653048845\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531316 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-config-data\") pod \"f94b9960-fa37-49f7-9fa7-0e01f855992b\" (UID: \"f94b9960-fa37-49f7-9fa7-0e01f855992b\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531348 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-httpd-run\") pod \"7209b8a6-5a97-4611-8183-84f1db25e2d1\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531383 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jnqv\" (UniqueName: \"kubernetes.io/projected/7209b8a6-5a97-4611-8183-84f1db25e2d1-kube-api-access-4jnqv\") pod \"7209b8a6-5a97-4611-8183-84f1db25e2d1\" (UID: \"7209b8a6-5a97-4611-8183-84f1db25e2d1\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531421 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9a8d972-b16f-421e-9fe4-f57653048845-etc-machine-id\") pod \"b9a8d972-b16f-421e-9fe4-f57653048845\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531441 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce23facc-1a0d-4c53-b696-b37a7525eba7-logs\") pod \"ce23facc-1a0d-4c53-b696-b37a7525eba7\" (UID: \"ce23facc-1a0d-4c53-b696-b37a7525eba7\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.531466 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-public-tls-certs\") pod \"b9a8d972-b16f-421e-9fe4-f57653048845\" (UID: \"b9a8d972-b16f-421e-9fe4-f57653048845\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.532145 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.532166 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvbj6\" (UniqueName: \"kubernetes.io/projected/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-kube-api-access-hvbj6\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.532188 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf68g\" (UniqueName: \"kubernetes.io/projected/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-api-access-qf68g\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.532200 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc12ffc5-e026-42a6-8ddc-43961c535787-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.532212 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70a59d46-95bf-4681-b61d-22239c638737-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.537708 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a8d972-b16f-421e-9fe4-f57653048845-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b9a8d972-b16f-421e-9fe4-f57653048845" (UID: "b9a8d972-b16f-421e-9fe4-f57653048845"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.537873 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7209b8a6-5a97-4611-8183-84f1db25e2d1" (UID: "7209b8a6-5a97-4611-8183-84f1db25e2d1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.538164 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f94b9960-fa37-49f7-9fa7-0e01f855992b" (UID: "f94b9960-fa37-49f7-9fa7-0e01f855992b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.540112 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9a8d972-b16f-421e-9fe4-f57653048845-logs" (OuterVolumeSpecName: "logs") pod "b9a8d972-b16f-421e-9fe4-f57653048845" (UID: "b9a8d972-b16f-421e-9fe4-f57653048845"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.540390 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70a59d46-95bf-4681-b61d-22239c638737-kube-api-access-gl4bh" (OuterVolumeSpecName: "kube-api-access-gl4bh") pod "70a59d46-95bf-4681-b61d-22239c638737" (UID: "70a59d46-95bf-4681-b61d-22239c638737"). InnerVolumeSpecName "kube-api-access-gl4bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.544814 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8d71630-681e-4126-bb64-e8b03f83ce97-kube-api-access-67gww" (OuterVolumeSpecName: "kube-api-access-67gww") pod "a8d71630-681e-4126-bb64-e8b03f83ce97" (UID: "a8d71630-681e-4126-bb64-e8b03f83ce97"). InnerVolumeSpecName "kube-api-access-67gww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.545661 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7209b8a6-5a97-4611-8183-84f1db25e2d1-kube-api-access-4jnqv" (OuterVolumeSpecName: "kube-api-access-4jnqv") pod "7209b8a6-5a97-4611-8183-84f1db25e2d1" (UID: "7209b8a6-5a97-4611-8183-84f1db25e2d1"). InnerVolumeSpecName "kube-api-access-4jnqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.545743 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc12ffc5-e026-42a6-8ddc-43961c535787-kube-api-access-zqqx7" (OuterVolumeSpecName: "kube-api-access-zqqx7") pod "fc12ffc5-e026-42a6-8ddc-43961c535787" (UID: "fc12ffc5-e026-42a6-8ddc-43961c535787"). InnerVolumeSpecName "kube-api-access-zqqx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.545808 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce23facc-1a0d-4c53-b696-b37a7525eba7-logs" (OuterVolumeSpecName: "logs") pod "ce23facc-1a0d-4c53-b696-b37a7525eba7" (UID: "ce23facc-1a0d-4c53-b696-b37a7525eba7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.547534 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-logs" (OuterVolumeSpecName: "logs") pod "f94b9960-fa37-49f7-9fa7-0e01f855992b" (UID: "f94b9960-fa37-49f7-9fa7-0e01f855992b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.547826 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-logs" (OuterVolumeSpecName: "logs") pod "7209b8a6-5a97-4611-8183-84f1db25e2d1" (UID: "7209b8a6-5a97-4611-8183-84f1db25e2d1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.566893 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-scripts" (OuterVolumeSpecName: "scripts") pod "f94b9960-fa37-49f7-9fa7-0e01f855992b" (UID: "f94b9960-fa37-49f7-9fa7-0e01f855992b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.572454 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a8d972-b16f-421e-9fe4-f57653048845-kube-api-access-xk7nh" (OuterVolumeSpecName: "kube-api-access-xk7nh") pod "b9a8d972-b16f-421e-9fe4-f57653048845" (UID: "b9a8d972-b16f-421e-9fe4-f57653048845"). InnerVolumeSpecName "kube-api-access-xk7nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.575235 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "f94b9960-fa37-49f7-9fa7-0e01f855992b" (UID: "f94b9960-fa37-49f7-9fa7-0e01f855992b"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.575432 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-scripts" (OuterVolumeSpecName: "scripts") pod "b9a8d972-b16f-421e-9fe4-f57653048845" (UID: "b9a8d972-b16f-421e-9fe4-f57653048845"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.575758 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f94b9960-fa37-49f7-9fa7-0e01f855992b-kube-api-access-9p56z" (OuterVolumeSpecName: "kube-api-access-9p56z") pod "f94b9960-fa37-49f7-9fa7-0e01f855992b" (UID: "f94b9960-fa37-49f7-9fa7-0e01f855992b"). InnerVolumeSpecName "kube-api-access-9p56z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.575747 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b9a8d972-b16f-421e-9fe4-f57653048845" (UID: "b9a8d972-b16f-421e-9fe4-f57653048845"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.575993 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce23facc-1a0d-4c53-b696-b37a7525eba7-kube-api-access-z9kp5" (OuterVolumeSpecName: "kube-api-access-z9kp5") pod "ce23facc-1a0d-4c53-b696-b37a7525eba7" (UID: "ce23facc-1a0d-4c53-b696-b37a7525eba7"). InnerVolumeSpecName "kube-api-access-z9kp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.578510 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "7209b8a6-5a97-4611-8183-84f1db25e2d1" (UID: "7209b8a6-5a97-4611-8183-84f1db25e2d1"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.601427 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-scripts" (OuterVolumeSpecName: "scripts") pod "7209b8a6-5a97-4611-8183-84f1db25e2d1" (UID: "7209b8a6-5a97-4611-8183-84f1db25e2d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.610774 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "d2e0b8db-ac91-432d-a87c-e741e441c33a" (UID: "d2e0b8db-ac91-432d-a87c-e741e441c33a"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634243 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634659 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p56z\" (UniqueName: \"kubernetes.io/projected/f94b9960-fa37-49f7-9fa7-0e01f855992b-kube-api-access-9p56z\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634676 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634688 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634704 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634717 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634729 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634741 4944 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634754 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9a8d972-b16f-421e-9fe4-f57653048845-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634767 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk7nh\" (UniqueName: \"kubernetes.io/projected/b9a8d972-b16f-421e-9fe4-f57653048845-kube-api-access-xk7nh\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634801 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634815 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9kp5\" (UniqueName: \"kubernetes.io/projected/ce23facc-1a0d-4c53-b696-b37a7525eba7-kube-api-access-z9kp5\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634827 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634840 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f94b9960-fa37-49f7-9fa7-0e01f855992b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634851 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7209b8a6-5a97-4611-8183-84f1db25e2d1-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634866 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jnqv\" (UniqueName: \"kubernetes.io/projected/7209b8a6-5a97-4611-8183-84f1db25e2d1-kube-api-access-4jnqv\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634879 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67gww\" (UniqueName: \"kubernetes.io/projected/a8d71630-681e-4126-bb64-e8b03f83ce97-kube-api-access-67gww\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634892 4944 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9a8d972-b16f-421e-9fe4-f57653048845-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634903 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce23facc-1a0d-4c53-b696-b37a7525eba7-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634915 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqqx7\" (UniqueName: \"kubernetes.io/projected/fc12ffc5-e026-42a6-8ddc-43961c535787-kube-api-access-zqqx7\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.634929 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl4bh\" (UniqueName: \"kubernetes.io/projected/70a59d46-95bf-4681-b61d-22239c638737-kube-api-access-gl4bh\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.639209 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-config-data" (OuterVolumeSpecName: "config-data") pod "a8d71630-681e-4126-bb64-e8b03f83ce97" (UID: "a8d71630-681e-4126-bb64-e8b03f83ce97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.639752 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder575d-account-delete-8dqqk" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.662379 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8d71630-681e-4126-bb64-e8b03f83ce97" (UID: "a8d71630-681e-4126-bb64-e8b03f83ce97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.664555 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.669247 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement0cdc-account-delete-sl7j9" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.698407 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance509f-account-delete-fx5bb" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.743535 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data-custom\") pod \"046a8c15-6acd-47fe-aef7-daac5085c2ec\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.743748 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-internal-tls-certs\") pod \"046a8c15-6acd-47fe-aef7-daac5085c2ec\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.743782 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/046a8c15-6acd-47fe-aef7-daac5085c2ec-logs\") pod \"046a8c15-6acd-47fe-aef7-daac5085c2ec\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.743855 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data\") pod \"046a8c15-6acd-47fe-aef7-daac5085c2ec\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.743900 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z75h\" (UniqueName: \"kubernetes.io/projected/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-kube-api-access-4z75h\") pod \"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2\" (UID: \"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.743919 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-combined-ca-bundle\") pod \"046a8c15-6acd-47fe-aef7-daac5085c2ec\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.743995 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-public-tls-certs\") pod \"046a8c15-6acd-47fe-aef7-daac5085c2ec\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.744060 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qs7r\" (UniqueName: \"kubernetes.io/projected/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-kube-api-access-5qs7r\") pod \"1db23d75-43f2-4bae-9e09-399d0ffb7ed5\" (UID: \"1db23d75-43f2-4bae-9e09-399d0ffb7ed5\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.744096 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtzf5\" (UniqueName: \"kubernetes.io/projected/69996228-3cfa-486e-8291-12e44f438c5b-kube-api-access-rtzf5\") pod \"69996228-3cfa-486e-8291-12e44f438c5b\" (UID: \"69996228-3cfa-486e-8291-12e44f438c5b\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.744134 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzczg\" (UniqueName: \"kubernetes.io/projected/046a8c15-6acd-47fe-aef7-daac5085c2ec-kube-api-access-dzczg\") pod \"046a8c15-6acd-47fe-aef7-daac5085c2ec\" (UID: \"046a8c15-6acd-47fe-aef7-daac5085c2ec\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.747467 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-operator-scripts\") pod \"1db23d75-43f2-4bae-9e09-399d0ffb7ed5\" (UID: \"1db23d75-43f2-4bae-9e09-399d0ffb7ed5\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.744432 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce23facc-1a0d-4c53-b696-b37a7525eba7" (UID: "ce23facc-1a0d-4c53-b696-b37a7525eba7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.746454 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc12ffc5-e026-42a6-8ddc-43961c535787" (UID: "fc12ffc5-e026-42a6-8ddc-43961c535787"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.747542 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-operator-scripts\") pod \"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2\" (UID: \"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.747594 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69996228-3cfa-486e-8291-12e44f438c5b-operator-scripts\") pod \"69996228-3cfa-486e-8291-12e44f438c5b\" (UID: \"69996228-3cfa-486e-8291-12e44f438c5b\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.748198 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/046a8c15-6acd-47fe-aef7-daac5085c2ec-logs" (OuterVolumeSpecName: "logs") pod "046a8c15-6acd-47fe-aef7-daac5085c2ec" (UID: "046a8c15-6acd-47fe-aef7-daac5085c2ec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.748311 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2b2d40c3-9522-46f4-85e7-fd724cb1bdf2" (UID: "2b2d40c3-9522-46f4-85e7-fd724cb1bdf2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.748626 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69996228-3cfa-486e-8291-12e44f438c5b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "69996228-3cfa-486e-8291-12e44f438c5b" (UID: "69996228-3cfa-486e-8291-12e44f438c5b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.748685 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1db23d75-43f2-4bae-9e09-399d0ffb7ed5" (UID: "1db23d75-43f2-4bae-9e09-399d0ffb7ed5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.748749 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/046a8c15-6acd-47fe-aef7-daac5085c2ec-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.748768 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.748787 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.748799 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.748811 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69996228-3cfa-486e-8291-12e44f438c5b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.748823 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.748836 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8d71630-681e-4126-bb64-e8b03f83ce97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.776793 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "046a8c15-6acd-47fe-aef7-daac5085c2ec" (UID: "046a8c15-6acd-47fe-aef7-daac5085c2ec"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.777583 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/046a8c15-6acd-47fe-aef7-daac5085c2ec-kube-api-access-dzczg" (OuterVolumeSpecName: "kube-api-access-dzczg") pod "046a8c15-6acd-47fe-aef7-daac5085c2ec" (UID: "046a8c15-6acd-47fe-aef7-daac5085c2ec"). InnerVolumeSpecName "kube-api-access-dzczg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.782036 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-kube-api-access-5qs7r" (OuterVolumeSpecName: "kube-api-access-5qs7r") pod "1db23d75-43f2-4bae-9e09-399d0ffb7ed5" (UID: "1db23d75-43f2-4bae-9e09-399d0ffb7ed5"). InnerVolumeSpecName "kube-api-access-5qs7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.787525 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-kube-api-access-4z75h" (OuterVolumeSpecName: "kube-api-access-4z75h") pod "2b2d40c3-9522-46f4-85e7-fd724cb1bdf2" (UID: "2b2d40c3-9522-46f4-85e7-fd724cb1bdf2"). InnerVolumeSpecName "kube-api-access-4z75h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.788155 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69996228-3cfa-486e-8291-12e44f438c5b-kube-api-access-rtzf5" (OuterVolumeSpecName: "kube-api-access-rtzf5") pod "69996228-3cfa-486e-8291-12e44f438c5b" (UID: "69996228-3cfa-486e-8291-12e44f438c5b"). InnerVolumeSpecName "kube-api-access-rtzf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.804194 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-config-data" (OuterVolumeSpecName: "config-data") pod "90fb3cc2-0725-48dc-9ad5-5c8da27930f8" (UID: "90fb3cc2-0725-48dc-9ad5-5c8da27930f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.804432 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-config-data" (OuterVolumeSpecName: "config-data") pod "fc12ffc5-e026-42a6-8ddc-43961c535787" (UID: "fc12ffc5-e026-42a6-8ddc-43961c535787"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.841721 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90fb3cc2-0725-48dc-9ad5-5c8da27930f8" (UID: "90fb3cc2-0725-48dc-9ad5-5c8da27930f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: E1124 09:14:28.851630 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.851663 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.851684 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90fb3cc2-0725-48dc-9ad5-5c8da27930f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.851750 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z75h\" (UniqueName: \"kubernetes.io/projected/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2-kube-api-access-4z75h\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.851762 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qs7r\" (UniqueName: \"kubernetes.io/projected/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-kube-api-access-5qs7r\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.851904 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtzf5\" (UniqueName: \"kubernetes.io/projected/69996228-3cfa-486e-8291-12e44f438c5b-kube-api-access-rtzf5\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.851917 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzczg\" (UniqueName: \"kubernetes.io/projected/046a8c15-6acd-47fe-aef7-daac5085c2ec-kube-api-access-dzczg\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.851929 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.851937 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1db23d75-43f2-4bae-9e09-399d0ffb7ed5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.851946 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: E1124 09:14:28.854333 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data podName:12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5 nodeName:}" failed. No retries permitted until 2025-11-24 09:14:36.854309755 +0000 UTC m=+1337.388750217 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data") pod "rabbitmq-cell1-server-0" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5") : configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.891630 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.891844 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapie0a2-account-delete-bvsxd" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.898568 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0eec8-account-delete-jtnrx" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.922397 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "d2e0b8db-ac91-432d-a87c-e741e441c33a" (UID: "d2e0b8db-ac91-432d-a87c-e741e441c33a"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.945015 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2e0b8db-ac91-432d-a87c-e741e441c33a" (UID: "d2e0b8db-ac91-432d-a87c-e741e441c33a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.953730 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-operator-scripts\") pod \"2a65aeee-1ac4-4ae3-b713-ad1d0074f797\" (UID: \"2a65aeee-1ac4-4ae3-b713-ad1d0074f797\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.953907 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndpl2\" (UniqueName: \"kubernetes.io/projected/b69c572d-53a7-4d39-a8fc-4152846915a5-kube-api-access-ndpl2\") pod \"b69c572d-53a7-4d39-a8fc-4152846915a5\" (UID: \"b69c572d-53a7-4d39-a8fc-4152846915a5\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.954075 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqxb6\" (UniqueName: \"kubernetes.io/projected/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-kube-api-access-jqxb6\") pod \"2a65aeee-1ac4-4ae3-b713-ad1d0074f797\" (UID: \"2a65aeee-1ac4-4ae3-b713-ad1d0074f797\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.954143 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b69c572d-53a7-4d39-a8fc-4152846915a5-operator-scripts\") pod \"b69c572d-53a7-4d39-a8fc-4152846915a5\" (UID: \"b69c572d-53a7-4d39-a8fc-4152846915a5\") " Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.954835 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.954861 4944 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.954875 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e0b8db-ac91-432d-a87c-e741e441c33a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.955354 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b69c572d-53a7-4d39-a8fc-4152846915a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b69c572d-53a7-4d39-a8fc-4152846915a5" (UID: "b69c572d-53a7-4d39-a8fc-4152846915a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.956216 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2a65aeee-1ac4-4ae3-b713-ad1d0074f797" (UID: "2a65aeee-1ac4-4ae3-b713-ad1d0074f797"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.965806 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-kube-api-access-jqxb6" (OuterVolumeSpecName: "kube-api-access-jqxb6") pod "2a65aeee-1ac4-4ae3-b713-ad1d0074f797" (UID: "2a65aeee-1ac4-4ae3-b713-ad1d0074f797"). InnerVolumeSpecName "kube-api-access-jqxb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.966408 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b69c572d-53a7-4d39-a8fc-4152846915a5-kube-api-access-ndpl2" (OuterVolumeSpecName: "kube-api-access-ndpl2") pod "b69c572d-53a7-4d39-a8fc-4152846915a5" (UID: "b69c572d-53a7-4d39-a8fc-4152846915a5"). InnerVolumeSpecName "kube-api-access-ndpl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.972027 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f94b9960-fa37-49f7-9fa7-0e01f855992b" (UID: "f94b9960-fa37-49f7-9fa7-0e01f855992b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.972415 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data" (OuterVolumeSpecName: "config-data") pod "046a8c15-6acd-47fe-aef7-daac5085c2ec" (UID: "046a8c15-6acd-47fe-aef7-daac5085c2ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.979162 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9a8d972-b16f-421e-9fe4-f57653048845" (UID: "b9a8d972-b16f-421e-9fe4-f57653048845"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:28 crc kubenswrapper[4944]: I1124 09:14:28.991897 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.007936 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-config-data" (OuterVolumeSpecName: "config-data") pod "7209b8a6-5a97-4611-8183-84f1db25e2d1" (UID: "7209b8a6-5a97-4611-8183-84f1db25e2d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.022448 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "046a8c15-6acd-47fe-aef7-daac5085c2ec" (UID: "046a8c15-6acd-47fe-aef7-daac5085c2ec"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.023338 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-config-data" (OuterVolumeSpecName: "config-data") pod "f94b9960-fa37-49f7-9fa7-0e01f855992b" (UID: "f94b9960-fa37-49f7-9fa7-0e01f855992b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.055861 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70a59d46-95bf-4681-b61d-22239c638737" (UID: "70a59d46-95bf-4681-b61d-22239c638737"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.056487 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-combined-ca-bundle\") pod \"70a59d46-95bf-4681-b61d-22239c638737\" (UID: \"70a59d46-95bf-4681-b61d-22239c638737\") " Nov 24 09:14:29 crc kubenswrapper[4944]: W1124 09:14:29.056723 4944 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/70a59d46-95bf-4681-b61d-22239c638737/volumes/kubernetes.io~secret/combined-ca-bundle Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.056749 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70a59d46-95bf-4681-b61d-22239c638737" (UID: "70a59d46-95bf-4681-b61d-22239c638737"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057187 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057211 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057224 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057237 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057249 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndpl2\" (UniqueName: \"kubernetes.io/projected/b69c572d-53a7-4d39-a8fc-4152846915a5-kube-api-access-ndpl2\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057260 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqxb6\" (UniqueName: \"kubernetes.io/projected/2a65aeee-1ac4-4ae3-b713-ad1d0074f797-kube-api-access-jqxb6\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057271 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057282 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b69c572d-53a7-4d39-a8fc-4152846915a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057292 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057304 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057314 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.057324 4944 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.065474 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-config-data" (OuterVolumeSpecName: "config-data") pod "ce23facc-1a0d-4c53-b696-b37a7525eba7" (UID: "ce23facc-1a0d-4c53-b696-b37a7525eba7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.091118 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data" (OuterVolumeSpecName: "config-data") pod "b9a8d972-b16f-421e-9fe4-f57653048845" (UID: "b9a8d972-b16f-421e-9fe4-f57653048845"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.102283 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "046a8c15-6acd-47fe-aef7-daac5085c2ec" (UID: "046a8c15-6acd-47fe-aef7-daac5085c2ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.109608 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7209b8a6-5a97-4611-8183-84f1db25e2d1" (UID: "7209b8a6-5a97-4611-8183-84f1db25e2d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.113257 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-config-data" (OuterVolumeSpecName: "config-data") pod "70a59d46-95bf-4681-b61d-22239c638737" (UID: "70a59d46-95bf-4681-b61d-22239c638737"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.114498 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ce23facc-1a0d-4c53-b696-b37a7525eba7" (UID: "ce23facc-1a0d-4c53-b696-b37a7525eba7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.119898 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b9a8d972-b16f-421e-9fe4-f57653048845" (UID: "b9a8d972-b16f-421e-9fe4-f57653048845"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.132227 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fc12ffc5-e026-42a6-8ddc-43961c535787" (UID: "fc12ffc5-e026-42a6-8ddc-43961c535787"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.137908 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ce23facc-1a0d-4c53-b696-b37a7525eba7" (UID: "ce23facc-1a0d-4c53-b696-b37a7525eba7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.147700 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b9a8d972-b16f-421e-9fe4-f57653048845" (UID: "b9a8d972-b16f-421e-9fe4-f57653048845"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.148617 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "046a8c15-6acd-47fe-aef7-daac5085c2ec" (UID: "046a8c15-6acd-47fe-aef7-daac5085c2ec"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.160943 4944 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.160991 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.161003 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.161014 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.161028 4944 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc12ffc5-e026-42a6-8ddc-43961c535787-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.161251 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.161262 4944 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.161273 4944 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce23facc-1a0d-4c53-b696-b37a7525eba7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.161283 4944 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.161293 4944 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9a8d972-b16f-421e-9fe4-f57653048845-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.161303 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/046a8c15-6acd-47fe-aef7-daac5085c2ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.167075 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f94b9960-fa37-49f7-9fa7-0e01f855992b" (UID: "f94b9960-fa37-49f7-9fa7-0e01f855992b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.179631 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "70a59d46-95bf-4681-b61d-22239c638737" (UID: "70a59d46-95bf-4681-b61d-22239c638737"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.184057 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7209b8a6-5a97-4611-8183-84f1db25e2d1" (UID: "7209b8a6-5a97-4611-8183-84f1db25e2d1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.209219 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapie0a2-account-delete-bvsxd" event={"ID":"2a65aeee-1ac4-4ae3-b713-ad1d0074f797","Type":"ContainerDied","Data":"be44c541c7e303f957c6328e9e73f5913708f7866af336580d068740794278c5"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.209264 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be44c541c7e303f957c6328e9e73f5913708f7866af336580d068740794278c5" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.209360 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapie0a2-account-delete-bvsxd" Nov 24 09:14:29 crc kubenswrapper[4944]: E1124 09:14:29.215128 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c is running failed: container process not found" containerID="5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 09:14:29 crc kubenswrapper[4944]: E1124 09:14:29.215407 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c is running failed: container process not found" containerID="5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 09:14:29 crc kubenswrapper[4944]: E1124 09:14:29.215888 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c is running failed: container process not found" containerID="5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 24 09:14:29 crc kubenswrapper[4944]: E1124 09:14:29.215916 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerName="ovn-northd" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.217253 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.217453 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9a8d972-b16f-421e-9fe4-f57653048845","Type":"ContainerDied","Data":"e6a761e29191681885e80ca6fb5d64dc5f3a481e473545b3ccdabff09f4c9d3a"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.217493 4944 scope.go:117] "RemoveContainer" containerID="84fadfabdc280facbbd9a3442d721a8b2b1c0298a2c5bd828ac25612a83c8d81" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.221220 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5ccb7b4f9b-mcdmd" event={"ID":"70a59d46-95bf-4681-b61d-22239c638737","Type":"ContainerDied","Data":"c3f8d5efd38123a3c9e348998d2cbf89144b5041be52f1cdb8f560f749d658c2"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.221303 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5ccb7b4f9b-mcdmd" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.232838 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "70a59d46-95bf-4681-b61d-22239c638737" (UID: "70a59d46-95bf-4681-b61d-22239c638737"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.237263 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2dfd2e16-b145-4464-8c5a-bff9a5908d45/ovn-northd/0.log" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.237320 4944 generic.go:334] "Generic (PLEG): container finished" podID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerID="5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c" exitCode=139 Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.237384 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2dfd2e16-b145-4464-8c5a-bff9a5908d45","Type":"ContainerDied","Data":"5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.237425 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2dfd2e16-b145-4464-8c5a-bff9a5908d45","Type":"ContainerDied","Data":"80a682e82fe06c1e1fb85d732a2b9c78e72af4a7b72dd32e5964e1b65bae10e7"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.237436 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80a682e82fe06c1e1fb85d732a2b9c78e72af4a7b72dd32e5964e1b65bae10e7" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.240537 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7209b8a6-5a97-4611-8183-84f1db25e2d1","Type":"ContainerDied","Data":"ddc83015cc841e6bff66ee193ae085c8026e3ab37954bbe93c399a4b0318c625"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.240568 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.245768 4944 generic.go:334] "Generic (PLEG): container finished" podID="35137b92-4f56-4399-8121-89bdb126162b" containerID="10473e2e58eff197e7d6054c78b1b43c310d54ae418a52fb6168c9c0b59133e8" exitCode=0 Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.245835 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"35137b92-4f56-4399-8121-89bdb126162b","Type":"ContainerDied","Data":"10473e2e58eff197e7d6054c78b1b43c310d54ae418a52fb6168c9c0b59133e8"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.245896 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"35137b92-4f56-4399-8121-89bdb126162b","Type":"ContainerDied","Data":"646c15a367bff50291e941fd08d418827b83f1e8278127d694dff9bd10c89f4a"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.245910 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="646c15a367bff50291e941fd08d418827b83f1e8278127d694dff9bd10c89f4a" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.248413 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement0cdc-account-delete-sl7j9" event={"ID":"69996228-3cfa-486e-8291-12e44f438c5b","Type":"ContainerDied","Data":"84d709e1a303c0866c88c4f8e95417f7ad9f12fe9bcb9c507763a9aae0a1ee4e"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.248460 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84d709e1a303c0866c88c4f8e95417f7ad9f12fe9bcb9c507763a9aae0a1ee4e" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.248483 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement0cdc-account-delete-sl7j9" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.260621 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance509f-account-delete-fx5bb" event={"ID":"1db23d75-43f2-4bae-9e09-399d0ffb7ed5","Type":"ContainerDied","Data":"96fe8c2872672ff999266afb3d7e5a06e9872d7521b5424a28a5b52a0902127c"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.260658 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96fe8c2872672ff999266afb3d7e5a06e9872d7521b5424a28a5b52a0902127c" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.260762 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance509f-account-delete-fx5bb" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.262504 4944 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7209b8a6-5a97-4611-8183-84f1db25e2d1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.262538 4944 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f94b9960-fa37-49f7-9fa7-0e01f855992b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.262554 4944 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.262567 4944 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70a59d46-95bf-4681-b61d-22239c638737-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.262566 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron7ac6-account-delete-7kth8" event={"ID":"11d8c8f9-ee83-4450-9a76-3c0a020652de","Type":"ContainerDied","Data":"8de6f9b2902383bc519a27bcb1dbae6df05a0319181cea964b80d672e04e2480"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.262589 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8de6f9b2902383bc519a27bcb1dbae6df05a0319181cea964b80d672e04e2480" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.268644 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc12ffc5-e026-42a6-8ddc-43961c535787","Type":"ContainerDied","Data":"6244c6b898aad9686ec3893884879e2cc73ae2fbb8dc7c6685421ec110803fa9"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.268664 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.275207 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8d71630-681e-4126-bb64-e8b03f83ce97","Type":"ContainerDied","Data":"27eb10684971e552edf6b8cd884f0217f8df9fbd74270aa468d962957d555e20"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.275334 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.283037 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f94b9960-fa37-49f7-9fa7-0e01f855992b","Type":"ContainerDied","Data":"7aed0246358a3d9b0b2e65a77b8c37b393752525ae5e977f142c01e137b22d55"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.283303 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.286906 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d2e0b8db-ac91-432d-a87c-e741e441c33a","Type":"ContainerDied","Data":"91a4015d43e34cead4d81b303d5fd64d6ce15ca8f651b251392b1550a7892360"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.287023 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.291787 4944 generic.go:334] "Generic (PLEG): container finished" podID="18d26ea6-eff1-4ad5-b50b-dd233a22e982" containerID="1a95527c06d2048bd5100167ef2d1b762657135708fcae5c22fd4931034b3057" exitCode=0 Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.291880 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"18d26ea6-eff1-4ad5-b50b-dd233a22e982","Type":"ContainerDied","Data":"1a95527c06d2048bd5100167ef2d1b762657135708fcae5c22fd4931034b3057"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.295311 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder575d-account-delete-8dqqk" event={"ID":"2b2d40c3-9522-46f4-85e7-fd724cb1bdf2","Type":"ContainerDied","Data":"5484dcdc1c11dd6f328aac9c91dc1ec5d3655a5649af2c5c9c971f820d6b01bc"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.295364 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5484dcdc1c11dd6f328aac9c91dc1ec5d3655a5649af2c5c9c971f820d6b01bc" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.295330 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder575d-account-delete-8dqqk" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.297810 4944 generic.go:334] "Generic (PLEG): container finished" podID="67e92d3f-3532-467f-bf38-c9c3107d4fcb" containerID="ea9e352d0f24eca87b4c56a85b5935690423ccc0712e37d2ea78fde15ddd4336" exitCode=0 Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.297898 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"67e92d3f-3532-467f-bf38-c9c3107d4fcb","Type":"ContainerDied","Data":"ea9e352d0f24eca87b4c56a85b5935690423ccc0712e37d2ea78fde15ddd4336"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.300172 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce23facc-1a0d-4c53-b696-b37a7525eba7","Type":"ContainerDied","Data":"f47f97b42ee843b2530865726437b567dd13721da1dc0af7e10f87abd25997a5"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.300300 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.302299 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f49db88bb-dzrb6" event={"ID":"046a8c15-6acd-47fe-aef7-daac5085c2ec","Type":"ContainerDied","Data":"01f911e69ae02d9b8cfae324442ea1d08c8cc067ff998c1912785fe2fe020c3c"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.302434 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f49db88bb-dzrb6" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.307601 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"90fb3cc2-0725-48dc-9ad5-5c8da27930f8","Type":"ContainerDied","Data":"28935e94026f788d13d8f73bdafe9c288381d3873af32802f62205a576bebfee"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.307756 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.313459 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0eec8-account-delete-jtnrx" event={"ID":"b69c572d-53a7-4d39-a8fc-4152846915a5","Type":"ContainerDied","Data":"11f5eb7d3d7c556ad26a6226db3ba58bc6b217dd6a967125a016d0a426fa34fd"} Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.313523 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0eec8-account-delete-jtnrx" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.465908 4944 scope.go:117] "RemoveContainer" containerID="ef2d54ce03a16a4e0d23c0030984dc6c28b947390104ad9f47f988b5741614d5" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.499189 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron7ac6-account-delete-7kth8" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.512433 4944 scope.go:117] "RemoveContainer" containerID="e42a48dc5f9a748005eb91765bdfa38b2ad4b4f974b03b15a72a6b00023e47a2" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.532035 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2dfd2e16-b145-4464-8c5a-bff9a5908d45/ovn-northd/0.log" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.532145 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.545805 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.558646 4944 scope.go:117] "RemoveContainer" containerID="4ad8bef4c684b16347a1d2ba8be599c596216230dd5dd3d9c7c090945c16c531" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.567455 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d8c8f9-ee83-4450-9a76-3c0a020652de-operator-scripts\") pod \"11d8c8f9-ee83-4450-9a76-3c0a020652de\" (UID: \"11d8c8f9-ee83-4450-9a76-3c0a020652de\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.567636 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95ckf\" (UniqueName: \"kubernetes.io/projected/11d8c8f9-ee83-4450-9a76-3c0a020652de-kube-api-access-95ckf\") pod \"11d8c8f9-ee83-4450-9a76-3c0a020652de\" (UID: \"11d8c8f9-ee83-4450-9a76-3c0a020652de\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.568257 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11d8c8f9-ee83-4450-9a76-3c0a020652de-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "11d8c8f9-ee83-4450-9a76-3c0a020652de" (UID: "11d8c8f9-ee83-4450-9a76-3c0a020652de"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.568628 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11d8c8f9-ee83-4450-9a76-3c0a020652de-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.592264 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11d8c8f9-ee83-4450-9a76-3c0a020652de-kube-api-access-95ckf" (OuterVolumeSpecName: "kube-api-access-95ckf") pod "11d8c8f9-ee83-4450-9a76-3c0a020652de" (UID: "11d8c8f9-ee83-4450-9a76-3c0a020652de"). InnerVolumeSpecName "kube-api-access-95ckf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.669481 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-config\") pod \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670328 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-config" (OuterVolumeSpecName: "config") pod "2dfd2e16-b145-4464-8c5a-bff9a5908d45" (UID: "2dfd2e16-b145-4464-8c5a-bff9a5908d45"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670480 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-rundir\") pod \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670572 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-scripts\") pod \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670650 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-northd-tls-certs\") pod \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670685 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-memcached-tls-certs\") pod \"35137b92-4f56-4399-8121-89bdb126162b\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670723 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-combined-ca-bundle\") pod \"35137b92-4f56-4399-8121-89bdb126162b\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670803 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z69d6\" (UniqueName: \"kubernetes.io/projected/2dfd2e16-b145-4464-8c5a-bff9a5908d45-kube-api-access-z69d6\") pod \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670848 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-combined-ca-bundle\") pod \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670860 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "2dfd2e16-b145-4464-8c5a-bff9a5908d45" (UID: "2dfd2e16-b145-4464-8c5a-bff9a5908d45"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670919 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-config-data\") pod \"35137b92-4f56-4399-8121-89bdb126162b\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670963 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jrgm\" (UniqueName: \"kubernetes.io/projected/35137b92-4f56-4399-8121-89bdb126162b-kube-api-access-8jrgm\") pod \"35137b92-4f56-4399-8121-89bdb126162b\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.670989 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-kolla-config\") pod \"35137b92-4f56-4399-8121-89bdb126162b\" (UID: \"35137b92-4f56-4399-8121-89bdb126162b\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.671016 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-metrics-certs-tls-certs\") pod \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\" (UID: \"2dfd2e16-b145-4464-8c5a-bff9a5908d45\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.671831 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-scripts" (OuterVolumeSpecName: "scripts") pod "2dfd2e16-b145-4464-8c5a-bff9a5908d45" (UID: "2dfd2e16-b145-4464-8c5a-bff9a5908d45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.672106 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.672125 4944 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.672139 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2dfd2e16-b145-4464-8c5a-bff9a5908d45-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.672152 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95ckf\" (UniqueName: \"kubernetes.io/projected/11d8c8f9-ee83-4450-9a76-3c0a020652de-kube-api-access-95ckf\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.672791 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-config-data" (OuterVolumeSpecName: "config-data") pod "35137b92-4f56-4399-8121-89bdb126162b" (UID: "35137b92-4f56-4399-8121-89bdb126162b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.675445 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "35137b92-4f56-4399-8121-89bdb126162b" (UID: "35137b92-4f56-4399-8121-89bdb126162b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.685609 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35137b92-4f56-4399-8121-89bdb126162b-kube-api-access-8jrgm" (OuterVolumeSpecName: "kube-api-access-8jrgm") pod "35137b92-4f56-4399-8121-89bdb126162b" (UID: "35137b92-4f56-4399-8121-89bdb126162b"). InnerVolumeSpecName "kube-api-access-8jrgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.710676 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dfd2e16-b145-4464-8c5a-bff9a5908d45-kube-api-access-z69d6" (OuterVolumeSpecName: "kube-api-access-z69d6") pod "2dfd2e16-b145-4464-8c5a-bff9a5908d45" (UID: "2dfd2e16-b145-4464-8c5a-bff9a5908d45"). InnerVolumeSpecName "kube-api-access-z69d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.751198 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35137b92-4f56-4399-8121-89bdb126162b" (UID: "35137b92-4f56-4399-8121-89bdb126162b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.775329 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.775364 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z69d6\" (UniqueName: \"kubernetes.io/projected/2dfd2e16-b145-4464-8c5a-bff9a5908d45-kube-api-access-z69d6\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.775376 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.775387 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jrgm\" (UniqueName: \"kubernetes.io/projected/35137b92-4f56-4399-8121-89bdb126162b-kube-api-access-8jrgm\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.775399 4944 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/35137b92-4f56-4399-8121-89bdb126162b-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.777951 4944 scope.go:117] "RemoveContainer" containerID="71ca2485422737d39ab2f15fc0ad24cd930f9cf353b7e908277d85b6a71d3f14" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.778628 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2dfd2e16-b145-4464-8c5a-bff9a5908d45" (UID: "2dfd2e16-b145-4464-8c5a-bff9a5908d45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.799483 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.810887 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "2dfd2e16-b145-4464-8c5a-bff9a5908d45" (UID: "2dfd2e16-b145-4464-8c5a-bff9a5908d45"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.812321 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.820648 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.839479 4944 scope.go:117] "RemoveContainer" containerID="b48283673309355d5ff56fb2136836b9670187973eb44ca2cc22b5681057bccf" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.842818 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.852329 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "2dfd2e16-b145-4464-8c5a-bff9a5908d45" (UID: "2dfd2e16-b145-4464-8c5a-bff9a5908d45"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.853092 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.858198 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "35137b92-4f56-4399-8121-89bdb126162b" (UID: "35137b92-4f56-4399-8121-89bdb126162b"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.870203 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878193 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878262 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/67e92d3f-3532-467f-bf38-c9c3107d4fcb-pod-info\") pod \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878292 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-erlang-cookie\") pod \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878333 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz227\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-kube-api-access-wz227\") pod \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878374 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-confd\") pod \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878393 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-operator-scripts\") pod \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878443 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-default\") pod \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878483 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-combined-ca-bundle\") pod \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878529 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data\") pod \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878573 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-generated\") pod \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878597 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-server-conf\") pod \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878616 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kolla-config\") pod \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878653 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-galera-tls-certs\") pod \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878681 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-plugins\") pod \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878718 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-tls\") pod \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878734 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsg45\" (UniqueName: \"kubernetes.io/projected/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kube-api-access-fsg45\") pod \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878774 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-plugins-conf\") pod \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878788 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\" (UID: \"18d26ea6-eff1-4ad5-b50b-dd233a22e982\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.878821 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/67e92d3f-3532-467f-bf38-c9c3107d4fcb-erlang-cookie-secret\") pod \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\" (UID: \"67e92d3f-3532-467f-bf38-c9c3107d4fcb\") " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.880241 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.880261 4944 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.880270 4944 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dfd2e16-b145-4464-8c5a-bff9a5908d45-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.880279 4944 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/35137b92-4f56-4399-8121-89bdb126162b-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.882501 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "67e92d3f-3532-467f-bf38-c9c3107d4fcb" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.882891 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "67e92d3f-3532-467f-bf38-c9c3107d4fcb" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.883520 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "18d26ea6-eff1-4ad5-b50b-dd233a22e982" (UID: "18d26ea6-eff1-4ad5-b50b-dd233a22e982"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.885606 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "67e92d3f-3532-467f-bf38-c9c3107d4fcb" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.888024 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "18d26ea6-eff1-4ad5-b50b-dd233a22e982" (UID: "18d26ea6-eff1-4ad5-b50b-dd233a22e982"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.888918 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.889401 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "18d26ea6-eff1-4ad5-b50b-dd233a22e982" (UID: "18d26ea6-eff1-4ad5-b50b-dd233a22e982"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.889478 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/67e92d3f-3532-467f-bf38-c9c3107d4fcb-pod-info" (OuterVolumeSpecName: "pod-info") pod "67e92d3f-3532-467f-bf38-c9c3107d4fcb" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.891094 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-kube-api-access-wz227" (OuterVolumeSpecName: "kube-api-access-wz227") pod "67e92d3f-3532-467f-bf38-c9c3107d4fcb" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb"). InnerVolumeSpecName "kube-api-access-wz227". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.891406 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "18d26ea6-eff1-4ad5-b50b-dd233a22e982" (UID: "18d26ea6-eff1-4ad5-b50b-dd233a22e982"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.892437 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kube-api-access-fsg45" (OuterVolumeSpecName: "kube-api-access-fsg45") pod "18d26ea6-eff1-4ad5-b50b-dd233a22e982" (UID: "18d26ea6-eff1-4ad5-b50b-dd233a22e982"). InnerVolumeSpecName "kube-api-access-fsg45". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.898171 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e92d3f-3532-467f-bf38-c9c3107d4fcb-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "67e92d3f-3532-467f-bf38-c9c3107d4fcb" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.919182 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.927003 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "67e92d3f-3532-467f-bf38-c9c3107d4fcb" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.941515 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "67e92d3f-3532-467f-bf38-c9c3107d4fcb" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.943296 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982443 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982470 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982480 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsg45\" (UniqueName: \"kubernetes.io/projected/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kube-api-access-fsg45\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982504 4944 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982514 4944 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/67e92d3f-3532-467f-bf38-c9c3107d4fcb-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982535 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982546 4944 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/67e92d3f-3532-467f-bf38-c9c3107d4fcb-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982559 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982571 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz227\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-kube-api-access-wz227\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982583 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982594 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982607 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/18d26ea6-eff1-4ad5-b50b-dd233a22e982-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.982620 4944 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/18d26ea6-eff1-4ad5-b50b-dd233a22e982-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.987014 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 09:14:29 crc kubenswrapper[4944]: I1124 09:14:29.996721 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0eec8-account-delete-jtnrx"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.003598 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.007144 4944 scope.go:117] "RemoveContainer" containerID="f4fda6a5047a5412abc68148082d7806e36494cd3997be68245586f663ec9996" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.010247 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0eec8-account-delete-jtnrx"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.011324 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "mysql-db") pod "18d26ea6-eff1-4ad5-b50b-dd233a22e982" (UID: "18d26ea6-eff1-4ad5-b50b-dd233a22e982"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.017166 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5ccb7b4f9b-mcdmd"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.046990 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-server-conf" (OuterVolumeSpecName: "server-conf") pod "67e92d3f-3532-467f-bf38-c9c3107d4fcb" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.051330 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-5ccb7b4f9b-mcdmd"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.058647 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18d26ea6-eff1-4ad5-b50b-dd233a22e982" (UID: "18d26ea6-eff1-4ad5-b50b-dd233a22e982"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.062724 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" podUID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.164:8080/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.062928 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-7bcd854fd7-z2d6k" podUID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.164:8080/healthcheck\": dial tcp 10.217.0.164:8080: i/o timeout" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.071645 4944 scope.go:117] "RemoveContainer" containerID="3dfe8ae924d36e9fb2dbf24bc72c97e4e845aaa92a56c64d0448011249204cfb" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.080950 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f49db88bb-dzrb6"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.085627 4944 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.085674 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.085684 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.085701 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.100163 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5f49db88bb-dzrb6"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.103585 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.104939 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data" (OuterVolumeSpecName: "config-data") pod "67e92d3f-3532-467f-bf38-c9c3107d4fcb" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.109934 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "18d26ea6-eff1-4ad5-b50b-dd233a22e982" (UID: "18d26ea6-eff1-4ad5-b50b-dd233a22e982"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.114610 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.132224 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.174192 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.190702 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.190742 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67e92d3f-3532-467f-bf38-c9c3107d4fcb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.190756 4944 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d26ea6-eff1-4ad5-b50b-dd233a22e982-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.194303 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.209768 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.216073 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.240399 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "67e92d3f-3532-467f-bf38-c9c3107d4fcb" (UID: "67e92d3f-3532-467f-bf38-c9c3107d4fcb"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.241340 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.247381 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.286909 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="046a8c15-6acd-47fe-aef7-daac5085c2ec" path="/var/lib/kubelet/pods/046a8c15-6acd-47fe-aef7-daac5085c2ec/volumes" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.287663 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70a59d46-95bf-4681-b61d-22239c638737" path="/var/lib/kubelet/pods/70a59d46-95bf-4681-b61d-22239c638737/volumes" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.288413 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7209b8a6-5a97-4611-8183-84f1db25e2d1" path="/var/lib/kubelet/pods/7209b8a6-5a97-4611-8183-84f1db25e2d1/volumes" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.289605 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90fb3cc2-0725-48dc-9ad5-5c8da27930f8" path="/var/lib/kubelet/pods/90fb3cc2-0725-48dc-9ad5-5c8da27930f8/volumes" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.290210 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8d71630-681e-4126-bb64-e8b03f83ce97" path="/var/lib/kubelet/pods/a8d71630-681e-4126-bb64-e8b03f83ce97/volumes" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.290765 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b69c572d-53a7-4d39-a8fc-4152846915a5" path="/var/lib/kubelet/pods/b69c572d-53a7-4d39-a8fc-4152846915a5/volumes" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.291905 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a8d972-b16f-421e-9fe4-f57653048845" path="/var/lib/kubelet/pods/b9a8d972-b16f-421e-9fe4-f57653048845/volumes" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.292364 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/67e92d3f-3532-467f-bf38-c9c3107d4fcb-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.292759 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce23facc-1a0d-4c53-b696-b37a7525eba7" path="/var/lib/kubelet/pods/ce23facc-1a0d-4c53-b696-b37a7525eba7/volumes" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.293377 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2e0b8db-ac91-432d-a87c-e741e441c33a" path="/var/lib/kubelet/pods/d2e0b8db-ac91-432d-a87c-e741e441c33a/volumes" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.294336 4944 scope.go:117] "RemoveContainer" containerID="2e79ba3b6da476e832ac8fa9f8ce25e21987e3f6fd5cbd139a2f9a79f5b464ef" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.295828 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f94b9960-fa37-49f7-9fa7-0e01f855992b" path="/var/lib/kubelet/pods/f94b9960-fa37-49f7-9fa7-0e01f855992b/volumes" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.296783 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" path="/var/lib/kubelet/pods/fc12ffc5-e026-42a6-8ddc-43961c535787/volumes" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.326244 4944 generic.go:334] "Generic (PLEG): container finished" podID="e3ea29cb-ae71-48af-8005-ba99c178a573" containerID="e3db2c95e6254ee19d5cd78241174e4c4f6100bc9626f271af3b94acb76058e6" exitCode=0 Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.326319 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" event={"ID":"e3ea29cb-ae71-48af-8005-ba99c178a573","Type":"ContainerDied","Data":"e3db2c95e6254ee19d5cd78241174e4c4f6100bc9626f271af3b94acb76058e6"} Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.334677 4944 generic.go:334] "Generic (PLEG): container finished" podID="d2495f8a-6f01-47e6-beff-f0d7345967a7" containerID="5950163d1a4608641fb113f50f6737d5436fc15138c0f58c9d9ec29a06eda578" exitCode=0 Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.334783 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5457b9857f-kg9gm" event={"ID":"d2495f8a-6f01-47e6-beff-f0d7345967a7","Type":"ContainerDied","Data":"5950163d1a4608641fb113f50f6737d5436fc15138c0f58c9d9ec29a06eda578"} Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.334839 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5457b9857f-kg9gm" event={"ID":"d2495f8a-6f01-47e6-beff-f0d7345967a7","Type":"ContainerDied","Data":"1ff67457d994ff3182e6c566098fc4eb751cdbf765fc343f619851795f6a8e7a"} Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.334855 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ff67457d994ff3182e6c566098fc4eb751cdbf765fc343f619851795f6a8e7a" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.337532 4944 generic.go:334] "Generic (PLEG): container finished" podID="3d9c665f-2be7-4b08-934d-b63036656b5f" containerID="9906021955dc6929c0ac49f16b105afc15f33b120aab600fc31ecf2026e5ba6e" exitCode=0 Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.337605 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-58f6cff88f-s86kk" event={"ID":"3d9c665f-2be7-4b08-934d-b63036656b5f","Type":"ContainerDied","Data":"9906021955dc6929c0ac49f16b105afc15f33b120aab600fc31ecf2026e5ba6e"} Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.343198 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"67e92d3f-3532-467f-bf38-c9c3107d4fcb","Type":"ContainerDied","Data":"31533e8484a6664f1f44ac64cf5fa6cf56f3bd282b02782b02cbcfbce6fe1aed"} Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.343247 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.347257 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"18d26ea6-eff1-4ad5-b50b-dd233a22e982","Type":"ContainerDied","Data":"5485cef22f5049df9ee59e90c4a019327f2dc0ec0af74eca05327bfcb5d47245"} Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.347716 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.351848 4944 generic.go:334] "Generic (PLEG): container finished" podID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerID="f1c621c73eb738098cb49670f9fd9711e1d1a037fa44d09a782be7678f0546b3" exitCode=0 Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.351959 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.352244 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4","Type":"ContainerDied","Data":"f1c621c73eb738098cb49670f9fd9711e1d1a037fa44d09a782be7678f0546b3"} Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.352326 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.352443 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron7ac6-account-delete-7kth8" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.365587 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.390689 4944 scope.go:117] "RemoveContainer" containerID="9a5211b2a8ad4ed0c93e5f29ec422b4395da812067c859ba887bf13db9c4afb6" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.436727 4944 scope.go:117] "RemoveContainer" containerID="95d6f710f056de96375b740c77544fefc875b760b049fb6331a5d9ea4510ffb4" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.439318 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.457470 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.475562 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.484241 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.490338 4944 scope.go:117] "RemoveContainer" containerID="cb293c82a7b8cd484e5348b25efb88a1b9008ec1eeee27738a2aeda8a4a18930" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.490958 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.495008 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-internal-tls-certs\") pod \"d2495f8a-6f01-47e6-beff-f0d7345967a7\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.495086 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-combined-ca-bundle\") pod \"d2495f8a-6f01-47e6-beff-f0d7345967a7\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.495167 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-credential-keys\") pod \"d2495f8a-6f01-47e6-beff-f0d7345967a7\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.495244 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-config-data\") pod \"d2495f8a-6f01-47e6-beff-f0d7345967a7\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.495301 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-public-tls-certs\") pod \"d2495f8a-6f01-47e6-beff-f0d7345967a7\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.495326 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-scripts\") pod \"d2495f8a-6f01-47e6-beff-f0d7345967a7\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.495382 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdx75\" (UniqueName: \"kubernetes.io/projected/d2495f8a-6f01-47e6-beff-f0d7345967a7-kube-api-access-wdx75\") pod \"d2495f8a-6f01-47e6-beff-f0d7345967a7\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.495421 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-fernet-keys\") pod \"d2495f8a-6f01-47e6-beff-f0d7345967a7\" (UID: \"d2495f8a-6f01-47e6-beff-f0d7345967a7\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.504243 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2495f8a-6f01-47e6-beff-f0d7345967a7-kube-api-access-wdx75" (OuterVolumeSpecName: "kube-api-access-wdx75") pod "d2495f8a-6f01-47e6-beff-f0d7345967a7" (UID: "d2495f8a-6f01-47e6-beff-f0d7345967a7"). InnerVolumeSpecName "kube-api-access-wdx75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.518125 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.518933 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-scripts" (OuterVolumeSpecName: "scripts") pod "d2495f8a-6f01-47e6-beff-f0d7345967a7" (UID: "d2495f8a-6f01-47e6-beff-f0d7345967a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.521708 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d2495f8a-6f01-47e6-beff-f0d7345967a7" (UID: "d2495f8a-6f01-47e6-beff-f0d7345967a7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.523845 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.530958 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d2495f8a-6f01-47e6-beff-f0d7345967a7" (UID: "d2495f8a-6f01-47e6-beff-f0d7345967a7"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.531329 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.540004 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.538890 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-config-data" (OuterVolumeSpecName: "config-data") pod "d2495f8a-6f01-47e6-beff-f0d7345967a7" (UID: "d2495f8a-6f01-47e6-beff-f0d7345967a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.550468 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d2495f8a-6f01-47e6-beff-f0d7345967a7" (UID: "d2495f8a-6f01-47e6-beff-f0d7345967a7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.559115 4944 scope.go:117] "RemoveContainer" containerID="1156a54f88c1c2dd0577bc2b082f7256b862e090c82e8172fc3df6b5bf9e5bdf" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.587387 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2495f8a-6f01-47e6-beff-f0d7345967a7" (UID: "d2495f8a-6f01-47e6-beff-f0d7345967a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.596489 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-sg-core-conf-yaml\") pod \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.596547 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-config-data\") pod \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.596573 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-combined-ca-bundle\") pod \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.596646 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-log-httpd\") pod \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.596680 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-ceilometer-tls-certs\") pod \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.596713 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-scripts\") pod \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.596778 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-run-httpd\") pod \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.596817 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhtzl\" (UniqueName: \"kubernetes.io/projected/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-kube-api-access-qhtzl\") pod \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\" (UID: \"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.597303 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.597328 4944 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.597343 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.597353 4944 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.597363 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.597391 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdx75\" (UniqueName: \"kubernetes.io/projected/d2495f8a-6f01-47e6-beff-f0d7345967a7-kube-api-access-wdx75\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.597426 4944 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.597451 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" (UID: "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.598536 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" (UID: "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.600248 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-kube-api-access-qhtzl" (OuterVolumeSpecName: "kube-api-access-qhtzl") pod "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" (UID: "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4"). InnerVolumeSpecName "kube-api-access-qhtzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.601803 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-scripts" (OuterVolumeSpecName: "scripts") pod "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" (UID: "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.609066 4944 scope.go:117] "RemoveContainer" containerID="ceef1855e41bb1981a2f6e513be92fa9f3b4e90878c9354637a799ca0b3bbbf4" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.629521 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" (UID: "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.641128 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.645772 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.678304 4944 scope.go:117] "RemoveContainer" containerID="7159a46717426d269e3e9ca766c99432598e2da70a9cf82e6a43abda53ec27b9" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.699664 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data\") pod \"e3ea29cb-ae71-48af-8005-ba99c178a573\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.700165 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-combined-ca-bundle\") pod \"3d9c665f-2be7-4b08-934d-b63036656b5f\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.700905 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ea29cb-ae71-48af-8005-ba99c178a573-logs\") pod \"e3ea29cb-ae71-48af-8005-ba99c178a573\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.701095 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d9c665f-2be7-4b08-934d-b63036656b5f-logs\") pod \"3d9c665f-2be7-4b08-934d-b63036656b5f\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.701489 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data-custom\") pod \"3d9c665f-2be7-4b08-934d-b63036656b5f\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.701507 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d9c665f-2be7-4b08-934d-b63036656b5f-logs" (OuterVolumeSpecName: "logs") pod "3d9c665f-2be7-4b08-934d-b63036656b5f" (UID: "3d9c665f-2be7-4b08-934d-b63036656b5f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.701979 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data\") pod \"3d9c665f-2be7-4b08-934d-b63036656b5f\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.702108 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9rxv\" (UniqueName: \"kubernetes.io/projected/3d9c665f-2be7-4b08-934d-b63036656b5f-kube-api-access-p9rxv\") pod \"3d9c665f-2be7-4b08-934d-b63036656b5f\" (UID: \"3d9c665f-2be7-4b08-934d-b63036656b5f\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.702235 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data-custom\") pod \"e3ea29cb-ae71-48af-8005-ba99c178a573\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.702294 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dznxr\" (UniqueName: \"kubernetes.io/projected/e3ea29cb-ae71-48af-8005-ba99c178a573-kube-api-access-dznxr\") pod \"e3ea29cb-ae71-48af-8005-ba99c178a573\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.702319 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-combined-ca-bundle\") pod \"e3ea29cb-ae71-48af-8005-ba99c178a573\" (UID: \"e3ea29cb-ae71-48af-8005-ba99c178a573\") " Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.702244 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3ea29cb-ae71-48af-8005-ba99c178a573-logs" (OuterVolumeSpecName: "logs") pod "e3ea29cb-ae71-48af-8005-ba99c178a573" (UID: "e3ea29cb-ae71-48af-8005-ba99c178a573"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.703625 4944 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.703647 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhtzl\" (UniqueName: \"kubernetes.io/projected/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-kube-api-access-qhtzl\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.703660 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ea29cb-ae71-48af-8005-ba99c178a573-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.703668 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d9c665f-2be7-4b08-934d-b63036656b5f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.703675 4944 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.703685 4944 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.703693 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.711173 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-xjbgk"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.712542 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3ea29cb-ae71-48af-8005-ba99c178a573-kube-api-access-dznxr" (OuterVolumeSpecName: "kube-api-access-dznxr") pod "e3ea29cb-ae71-48af-8005-ba99c178a573" (UID: "e3ea29cb-ae71-48af-8005-ba99c178a573"). InnerVolumeSpecName "kube-api-access-dznxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.713186 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d9c665f-2be7-4b08-934d-b63036656b5f-kube-api-access-p9rxv" (OuterVolumeSpecName: "kube-api-access-p9rxv") pod "3d9c665f-2be7-4b08-934d-b63036656b5f" (UID: "3d9c665f-2be7-4b08-934d-b63036656b5f"). InnerVolumeSpecName "kube-api-access-p9rxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.713326 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e3ea29cb-ae71-48af-8005-ba99c178a573" (UID: "e3ea29cb-ae71-48af-8005-ba99c178a573"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.713682 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3d9c665f-2be7-4b08-934d-b63036656b5f" (UID: "3d9c665f-2be7-4b08-934d-b63036656b5f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.715880 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" (UID: "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.716333 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d2495f8a-6f01-47e6-beff-f0d7345967a7" (UID: "d2495f8a-6f01-47e6-beff-f0d7345967a7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.718769 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-xjbgk"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.719757 4944 scope.go:117] "RemoveContainer" containerID="7aa8ffc64bce11fa46956c14a0b6ed75513f2964c1dd926a5e6be5120e48e88e" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.727632 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicand270-account-delete-h2tcc"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.734773 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-d270-account-create-trwxh"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.736494 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" (UID: "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.741175 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-d270-account-create-trwxh"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.750950 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d9c665f-2be7-4b08-934d-b63036656b5f" (UID: "3d9c665f-2be7-4b08-934d-b63036656b5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.754651 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3ea29cb-ae71-48af-8005-ba99c178a573" (UID: "e3ea29cb-ae71-48af-8005-ba99c178a573"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.754680 4944 scope.go:117] "RemoveContainer" containerID="58794c2902c7c98e7b2d3be65f6f669e114b85ae8de8b4c493118b02ba26a249" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.755421 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-config-data" (OuterVolumeSpecName: "config-data") pod "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" (UID: "5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.756303 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data" (OuterVolumeSpecName: "config-data") pod "e3ea29cb-ae71-48af-8005-ba99c178a573" (UID: "e3ea29cb-ae71-48af-8005-ba99c178a573"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.763507 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicand270-account-delete-h2tcc"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.775571 4944 scope.go:117] "RemoveContainer" containerID="0e567cd20074605f8fc1922dc6cc36a075eb716e38a58a3ff08e0c19da6b0f70" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.794905 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data" (OuterVolumeSpecName: "config-data") pod "3d9c665f-2be7-4b08-934d-b63036656b5f" (UID: "3d9c665f-2be7-4b08-934d-b63036656b5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.801483 4944 scope.go:117] "RemoveContainer" containerID="ea9e352d0f24eca87b4c56a85b5935690423ccc0712e37d2ea78fde15ddd4336" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805444 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805473 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805484 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805493 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9rxv\" (UniqueName: \"kubernetes.io/projected/3d9c665f-2be7-4b08-934d-b63036656b5f-kube-api-access-p9rxv\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805502 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805510 4944 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805519 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dznxr\" (UniqueName: \"kubernetes.io/projected/e3ea29cb-ae71-48af-8005-ba99c178a573-kube-api-access-dznxr\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805527 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805534 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ea29cb-ae71-48af-8005-ba99c178a573-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805541 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805550 4944 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2495f8a-6f01-47e6-beff-f0d7345967a7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.805560 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d9c665f-2be7-4b08-934d-b63036656b5f-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.854799 4944 scope.go:117] "RemoveContainer" containerID="49cd4f1b400eddca6bdd6df980eb6336eb7f945c92b13a7f0c34e97beb31f078" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.859258 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-b8fcn"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.871223 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-b8fcn"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.883349 4944 scope.go:117] "RemoveContainer" containerID="1a95527c06d2048bd5100167ef2d1b762657135708fcae5c22fd4931034b3057" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.895084 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-575d-account-create-c8gbm"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.902830 4944 scope.go:117] "RemoveContainer" containerID="f55d411db821167a7793432c6ee6466117e2082720794f9b45e02915d1b9298e" Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.905207 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder575d-account-delete-8dqqk"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.911684 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder575d-account-delete-8dqqk"] Nov 24 09:14:30 crc kubenswrapper[4944]: I1124 09:14:30.917370 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-575d-account-create-c8gbm"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.091326 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-bgpkk"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.125126 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-bgpkk"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.133174 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-509f-account-create-6wsxt"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.142202 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance509f-account-delete-fx5bb"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.153728 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance509f-account-delete-fx5bb"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.161751 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-509f-account-create-6wsxt"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.193091 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-d4j84"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.207120 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-d4j84"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.218225 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-0cdc-account-create-p4c9q"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.228104 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement0cdc-account-delete-sl7j9"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.235157 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-0cdc-account-create-p4c9q"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.245176 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement0cdc-account-delete-sl7j9"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.311812 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-97djh"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.320561 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-97djh"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.331226 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron7ac6-account-delete-7kth8"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.335364 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7ac6-account-create-t66tz"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.340618 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7ac6-account-create-t66tz"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.346226 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron7ac6-account-delete-7kth8"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.365115 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" event={"ID":"e3ea29cb-ae71-48af-8005-ba99c178a573","Type":"ContainerDied","Data":"9ff6376fa9603b040179576baa880c926be1eeeff20bc399721dcd84b1c41fe8"} Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.365199 4944 scope.go:117] "RemoveContainer" containerID="e3db2c95e6254ee19d5cd78241174e4c4f6100bc9626f271af3b94acb76058e6" Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.365442 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7b47cdb5f8-g5n25" Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.381805 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-58f6cff88f-s86kk" event={"ID":"3d9c665f-2be7-4b08-934d-b63036656b5f","Type":"ContainerDied","Data":"eb0ffeae65b43b2ec46ada236b5d55d28eb59b258a8954c9cfafb8df9eb31e28"} Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.382230 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-58f6cff88f-s86kk" Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.403845 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7b47cdb5f8-g5n25"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.410465 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-7b47cdb5f8-g5n25"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.443276 4944 scope.go:117] "RemoveContainer" containerID="5e2bf771bd1be4b28e3fa735cec21fcbbccd1e4807f6ac424dd4405dc8f8a329" Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.449794 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.450568 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4","Type":"ContainerDied","Data":"10020383316a643a77b2528f630a5557d012ead2d5b2ee865f54416fbdcbc49c"} Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.456591 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-58f6cff88f-s86kk"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.463125 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5457b9857f-kg9gm" Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.466445 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-58f6cff88f-s86kk"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.493168 4944 scope.go:117] "RemoveContainer" containerID="9906021955dc6929c0ac49f16b105afc15f33b120aab600fc31ecf2026e5ba6e" Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.494033 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-5bmbf"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.507897 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-5bmbf"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.515398 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapie0a2-account-delete-bvsxd"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.527011 4944 scope.go:117] "RemoveContainer" containerID="219c63bcbbda1c8ee292c94dcc9e2f39b33c5aab0d972baaee0d1dc1f454692d" Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.529532 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-e0a2-account-create-xss7d"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.541585 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapie0a2-account-delete-bvsxd"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.549572 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-e0a2-account-create-xss7d"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.562700 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.564102 4944 scope.go:117] "RemoveContainer" containerID="a84b239223662f1225b37375264550577380c4e7d01f7c269035f7022dd18d82" Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.570558 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.579424 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5457b9857f-kg9gm"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.587834 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5457b9857f-kg9gm"] Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.592063 4944 scope.go:117] "RemoveContainer" containerID="dbee395d02885c9a06d22bd217f37d54477844b3e398fa6029fcb08eebb7be4e" Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.615107 4944 scope.go:117] "RemoveContainer" containerID="f1c621c73eb738098cb49670f9fd9711e1d1a037fa44d09a782be7678f0546b3" Nov 24 09:14:31 crc kubenswrapper[4944]: I1124 09:14:31.638231 4944 scope.go:117] "RemoveContainer" containerID="73095b06c9c5460d64ea3a6e6ad293c16acb97dedadc392b482d5b45673686f5" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.286028 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f8b9684-d40c-4c61-aff8-59008e970331" path="/var/lib/kubelet/pods/0f8b9684-d40c-4c61-aff8-59008e970331/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.287193 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11d8c8f9-ee83-4450-9a76-3c0a020652de" path="/var/lib/kubelet/pods/11d8c8f9-ee83-4450-9a76-3c0a020652de/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.287921 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18d26ea6-eff1-4ad5-b50b-dd233a22e982" path="/var/lib/kubelet/pods/18d26ea6-eff1-4ad5-b50b-dd233a22e982/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.289218 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1db23d75-43f2-4bae-9e09-399d0ffb7ed5" path="/var/lib/kubelet/pods/1db23d75-43f2-4bae-9e09-399d0ffb7ed5/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.289784 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ef47237-852b-485d-954b-285d07f44af3" path="/var/lib/kubelet/pods/1ef47237-852b-485d-954b-285d07f44af3/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.290330 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a65aeee-1ac4-4ae3-b713-ad1d0074f797" path="/var/lib/kubelet/pods/2a65aeee-1ac4-4ae3-b713-ad1d0074f797/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.291449 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b2d40c3-9522-46f4-85e7-fd724cb1bdf2" path="/var/lib/kubelet/pods/2b2d40c3-9522-46f4-85e7-fd724cb1bdf2/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.292037 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" path="/var/lib/kubelet/pods/2dfd2e16-b145-4464-8c5a-bff9a5908d45/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.292693 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35137b92-4f56-4399-8121-89bdb126162b" path="/var/lib/kubelet/pods/35137b92-4f56-4399-8121-89bdb126162b/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.293869 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d9c665f-2be7-4b08-934d-b63036656b5f" path="/var/lib/kubelet/pods/3d9c665f-2be7-4b08-934d-b63036656b5f/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.294661 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d176dce-ca46-4a21-801e-42055ae16d1d" path="/var/lib/kubelet/pods/4d176dce-ca46-4a21-801e-42055ae16d1d/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.295285 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55a9098c-7257-46f7-888c-c4bb6f9421a5" path="/var/lib/kubelet/pods/55a9098c-7257-46f7-888c-c4bb6f9421a5/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.296344 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" path="/var/lib/kubelet/pods/5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.297266 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f08ed10-9454-4781-8a6b-000100a436dc" path="/var/lib/kubelet/pods/5f08ed10-9454-4781-8a6b-000100a436dc/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.297969 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67e92d3f-3532-467f-bf38-c9c3107d4fcb" path="/var/lib/kubelet/pods/67e92d3f-3532-467f-bf38-c9c3107d4fcb/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.298966 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69996228-3cfa-486e-8291-12e44f438c5b" path="/var/lib/kubelet/pods/69996228-3cfa-486e-8291-12e44f438c5b/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.299546 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a5dfba8-039e-4bcd-ab5b-f654731ecfb4" path="/var/lib/kubelet/pods/6a5dfba8-039e-4bcd-ab5b-f654731ecfb4/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.300111 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="787b579d-d704-469d-b0b9-c903c0eef0f9" path="/var/lib/kubelet/pods/787b579d-d704-469d-b0b9-c903c0eef0f9/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.301236 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dc3bfd5-b619-4b27-842a-5c5d47375c6a" path="/var/lib/kubelet/pods/8dc3bfd5-b619-4b27-842a-5c5d47375c6a/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: E1124 09:14:32.301259 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:32 crc kubenswrapper[4944]: E1124 09:14:32.301754 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.301937 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3235d75-d12c-4656-adb4-d0d485f2b45b" path="/var/lib/kubelet/pods/a3235d75-d12c-4656-adb4-d0d485f2b45b/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: E1124 09:14:32.302058 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:32 crc kubenswrapper[4944]: E1124 09:14:32.302100 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.302583 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1ea10a0-65c3-465c-bf40-e1804f588745" path="/var/lib/kubelet/pods/d1ea10a0-65c3-465c-bf40-e1804f588745/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: E1124 09:14:32.302696 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.303648 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2495f8a-6f01-47e6-beff-f0d7345967a7" path="/var/lib/kubelet/pods/d2495f8a-6f01-47e6-beff-f0d7345967a7/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.304876 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9e4a45c-7aa7-4c04-8c86-a7927a7ada80" path="/var/lib/kubelet/pods/d9e4a45c-7aa7-4c04-8c86-a7927a7ada80/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.305755 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc7dc25c-e0b6-4dc8-92f4-975dd74346a0" path="/var/lib/kubelet/pods/dc7dc25c-e0b6-4dc8-92f4-975dd74346a0/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.306425 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3ea29cb-ae71-48af-8005-ba99c178a573" path="/var/lib/kubelet/pods/e3ea29cb-ae71-48af-8005-ba99c178a573/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: E1124 09:14:32.306798 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:32 crc kubenswrapper[4944]: I1124 09:14:32.307900 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7851f98-142f-4123-be3b-c811c294954e" path="/var/lib/kubelet/pods/e7851f98-142f-4123-be3b-c811c294954e/volumes" Nov 24 09:14:32 crc kubenswrapper[4944]: E1124 09:14:32.308231 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:32 crc kubenswrapper[4944]: E1124 09:14:32.308286 4944 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovs-vswitchd" Nov 24 09:14:33 crc kubenswrapper[4944]: I1124 09:14:33.339425 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 09:14:33 crc kubenswrapper[4944]: I1124 09:14:33.340452 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": dial tcp 10.217.0.200:8775: i/o timeout" Nov 24 09:14:34 crc kubenswrapper[4944]: I1124 09:14:34.029321 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-5d5d955fb7-kf4q4" podUID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.151:9696/\": dial tcp 10.217.0.151:9696: connect: connection refused" Nov 24 09:14:34 crc kubenswrapper[4944]: I1124 09:14:34.736607 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="67e92d3f-3532-467f-bf38-c9c3107d4fcb" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: i/o timeout" Nov 24 09:14:36 crc kubenswrapper[4944]: E1124 09:14:36.923947 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:36 crc kubenswrapper[4944]: E1124 09:14:36.924401 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data podName:12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5 nodeName:}" failed. No retries permitted until 2025-11-24 09:14:52.92437366 +0000 UTC m=+1353.458814122 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data") pod "rabbitmq-cell1-server-0" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5") : configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:37 crc kubenswrapper[4944]: E1124 09:14:37.293471 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:37 crc kubenswrapper[4944]: E1124 09:14:37.294010 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:37 crc kubenswrapper[4944]: E1124 09:14:37.294461 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:37 crc kubenswrapper[4944]: E1124 09:14:37.294511 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server" Nov 24 09:14:37 crc kubenswrapper[4944]: E1124 09:14:37.297349 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:37 crc kubenswrapper[4944]: E1124 09:14:37.298690 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:37 crc kubenswrapper[4944]: E1124 09:14:37.299969 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:37 crc kubenswrapper[4944]: E1124 09:14:37.300032 4944 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovs-vswitchd" Nov 24 09:14:42 crc kubenswrapper[4944]: E1124 09:14:42.293783 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:42 crc kubenswrapper[4944]: E1124 09:14:42.294675 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:42 crc kubenswrapper[4944]: E1124 09:14:42.294868 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:42 crc kubenswrapper[4944]: E1124 09:14:42.294965 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:42 crc kubenswrapper[4944]: E1124 09:14:42.295002 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server" Nov 24 09:14:42 crc kubenswrapper[4944]: E1124 09:14:42.296737 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:42 crc kubenswrapper[4944]: E1124 09:14:42.298008 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:42 crc kubenswrapper[4944]: E1124 09:14:42.298062 4944 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovs-vswitchd" Nov 24 09:14:47 crc kubenswrapper[4944]: E1124 09:14:47.293619 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:47 crc kubenswrapper[4944]: E1124 09:14:47.294187 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:47 crc kubenswrapper[4944]: E1124 09:14:47.294395 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:47 crc kubenswrapper[4944]: E1124 09:14:47.294836 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 24 09:14:47 crc kubenswrapper[4944]: E1124 09:14:47.294885 4944 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server" Nov 24 09:14:47 crc kubenswrapper[4944]: E1124 09:14:47.295280 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:47 crc kubenswrapper[4944]: E1124 09:14:47.296546 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 24 09:14:47 crc kubenswrapper[4944]: E1124 09:14:47.296603 4944 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7gtjj" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovs-vswitchd" Nov 24 09:14:51 crc kubenswrapper[4944]: E1124 09:14:51.373993 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1d0474c_20bb_4653_8ce7_176bf4bc5f0b.slice/crio-conmon-c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1d0474c_20bb_4653_8ce7_176bf4bc5f0b.slice/crio-c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac.scope\": RecentStats: unable to find data in memory cache]" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.653459 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7gtjj_e1d0474c-20bb-4653-8ce7-176bf4bc5f0b/ovs-vswitchd/0.log" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.655001 4944 generic.go:334] "Generic (PLEG): container finished" podID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" exitCode=137 Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.655140 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7gtjj" event={"ID":"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b","Type":"ContainerDied","Data":"c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac"} Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.655177 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7gtjj" event={"ID":"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b","Type":"ContainerDied","Data":"6ec02dfe04c23880d4b3c99aa2deaa2b332f08aad186936b4a51d23a76d9db5d"} Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.655206 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ec02dfe04c23880d4b3c99aa2deaa2b332f08aad186936b4a51d23a76d9db5d" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.701977 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7gtjj_e1d0474c-20bb-4653-8ce7-176bf4bc5f0b/ovs-vswitchd/0.log" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.702811 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.774867 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-etc-ovs\") pod \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.774915 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9htrv\" (UniqueName: \"kubernetes.io/projected/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-kube-api-access-9htrv\") pod \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.774972 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-log\") pod \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.775021 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-run\") pod \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.775099 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-lib\") pod \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.775159 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-scripts\") pod \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\" (UID: \"e1d0474c-20bb-4653-8ce7-176bf4bc5f0b\") " Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.776470 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-scripts" (OuterVolumeSpecName: "scripts") pod "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" (UID: "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.776514 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" (UID: "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.776998 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-run" (OuterVolumeSpecName: "var-run") pod "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" (UID: "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.777041 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-lib" (OuterVolumeSpecName: "var-lib") pod "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" (UID: "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.777069 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-log" (OuterVolumeSpecName: "var-log") pod "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" (UID: "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.787617 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-kube-api-access-9htrv" (OuterVolumeSpecName: "kube-api-access-9htrv") pod "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" (UID: "e1d0474c-20bb-4653-8ce7-176bf4bc5f0b"). InnerVolumeSpecName "kube-api-access-9htrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.877490 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9htrv\" (UniqueName: \"kubernetes.io/projected/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-kube-api-access-9htrv\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.877537 4944 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-etc-ovs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.877550 4944 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-log\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.877560 4944 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.877571 4944 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-var-lib\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:51 crc kubenswrapper[4944]: I1124 09:14:51.877581 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.311641 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.385321 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data-custom\") pod \"53d02d15-b171-4e78-b528-9bec6362bf70\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.385496 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-scripts\") pod \"53d02d15-b171-4e78-b528-9bec6362bf70\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.385565 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-combined-ca-bundle\") pod \"53d02d15-b171-4e78-b528-9bec6362bf70\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.385609 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk6fn\" (UniqueName: \"kubernetes.io/projected/53d02d15-b171-4e78-b528-9bec6362bf70-kube-api-access-nk6fn\") pod \"53d02d15-b171-4e78-b528-9bec6362bf70\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.385646 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53d02d15-b171-4e78-b528-9bec6362bf70-etc-machine-id\") pod \"53d02d15-b171-4e78-b528-9bec6362bf70\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.385723 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data\") pod \"53d02d15-b171-4e78-b528-9bec6362bf70\" (UID: \"53d02d15-b171-4e78-b528-9bec6362bf70\") " Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.386461 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d02d15-b171-4e78-b528-9bec6362bf70-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "53d02d15-b171-4e78-b528-9bec6362bf70" (UID: "53d02d15-b171-4e78-b528-9bec6362bf70"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.386717 4944 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53d02d15-b171-4e78-b528-9bec6362bf70-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.389910 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d02d15-b171-4e78-b528-9bec6362bf70-kube-api-access-nk6fn" (OuterVolumeSpecName: "kube-api-access-nk6fn") pod "53d02d15-b171-4e78-b528-9bec6362bf70" (UID: "53d02d15-b171-4e78-b528-9bec6362bf70"). InnerVolumeSpecName "kube-api-access-nk6fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.390345 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "53d02d15-b171-4e78-b528-9bec6362bf70" (UID: "53d02d15-b171-4e78-b528-9bec6362bf70"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.392086 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-scripts" (OuterVolumeSpecName: "scripts") pod "53d02d15-b171-4e78-b528-9bec6362bf70" (UID: "53d02d15-b171-4e78-b528-9bec6362bf70"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.431560 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53d02d15-b171-4e78-b528-9bec6362bf70" (UID: "53d02d15-b171-4e78-b528-9bec6362bf70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.488413 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.488471 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.488486 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.488500 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk6fn\" (UniqueName: \"kubernetes.io/projected/53d02d15-b171-4e78-b528-9bec6362bf70-kube-api-access-nk6fn\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.503796 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data" (OuterVolumeSpecName: "config-data") pod "53d02d15-b171-4e78-b528-9bec6362bf70" (UID: "53d02d15-b171-4e78-b528-9bec6362bf70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.590364 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d02d15-b171-4e78-b528-9bec6362bf70-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.670419 4944 generic.go:334] "Generic (PLEG): container finished" podID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerID="d6d5476d01856b9678b39efa998de2c30d0cc6772d172d0bf1aa500eb8ebaca7" exitCode=137 Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.670487 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"d6d5476d01856b9678b39efa998de2c30d0cc6772d172d0bf1aa500eb8ebaca7"} Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.674846 4944 generic.go:334] "Generic (PLEG): container finished" podID="53d02d15-b171-4e78-b528-9bec6362bf70" containerID="b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b" exitCode=137 Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.674975 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7gtjj" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.675786 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.677446 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"53d02d15-b171-4e78-b528-9bec6362bf70","Type":"ContainerDied","Data":"b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b"} Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.677514 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"53d02d15-b171-4e78-b528-9bec6362bf70","Type":"ContainerDied","Data":"8a40812f7cb63d0cb175cee00130c9703f7a8c995b196814512aaf1b5b9db410"} Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.677538 4944 scope.go:117] "RemoveContainer" containerID="fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.713953 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-7gtjj"] Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.719229 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-7gtjj"] Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.719314 4944 scope.go:117] "RemoveContainer" containerID="b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.723476 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.735632 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.752158 4944 scope.go:117] "RemoveContainer" containerID="fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f" Nov 24 09:14:52 crc kubenswrapper[4944]: E1124 09:14:52.752699 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f\": container with ID starting with fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f not found: ID does not exist" containerID="fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.752751 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f"} err="failed to get container status \"fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f\": rpc error: code = NotFound desc = could not find container \"fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f\": container with ID starting with fb0c39821dab6691335c9029c8f54771f2d2b162e68078e37bbb31883dcfa99f not found: ID does not exist" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.752779 4944 scope.go:117] "RemoveContainer" containerID="b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b" Nov 24 09:14:52 crc kubenswrapper[4944]: E1124 09:14:52.753350 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b\": container with ID starting with b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b not found: ID does not exist" containerID="b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.753373 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b"} err="failed to get container status \"b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b\": rpc error: code = NotFound desc = could not find container \"b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b\": container with ID starting with b3716817d4dd9a38ac6965e9853aa5f86a675f2715914ddfa2cf2d80bd8e689b not found: ID does not exist" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.856778 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.996797 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-lock\") pod \"6918e56a-00f4-4f99-b3ef-c65be06c428a\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.996902 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-cache\") pod \"6918e56a-00f4-4f99-b3ef-c65be06c428a\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.997082 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift\") pod \"6918e56a-00f4-4f99-b3ef-c65be06c428a\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.997150 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"6918e56a-00f4-4f99-b3ef-c65be06c428a\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.997187 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxs7n\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-kube-api-access-nxs7n\") pod \"6918e56a-00f4-4f99-b3ef-c65be06c428a\" (UID: \"6918e56a-00f4-4f99-b3ef-c65be06c428a\") " Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.997493 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-lock" (OuterVolumeSpecName: "lock") pod "6918e56a-00f4-4f99-b3ef-c65be06c428a" (UID: "6918e56a-00f4-4f99-b3ef-c65be06c428a"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:52 crc kubenswrapper[4944]: E1124 09:14:52.997572 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:52 crc kubenswrapper[4944]: E1124 09:14:52.997619 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data podName:12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5 nodeName:}" failed. No retries permitted until 2025-11-24 09:15:24.99760554 +0000 UTC m=+1385.532046002 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data") pod "rabbitmq-cell1-server-0" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5") : configmap "rabbitmq-cell1-config-data" not found Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.997678 4944 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-lock\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:52 crc kubenswrapper[4944]: I1124 09:14:52.997667 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-cache" (OuterVolumeSpecName: "cache") pod "6918e56a-00f4-4f99-b3ef-c65be06c428a" (UID: "6918e56a-00f4-4f99-b3ef-c65be06c428a"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.002297 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "swift") pod "6918e56a-00f4-4f99-b3ef-c65be06c428a" (UID: "6918e56a-00f4-4f99-b3ef-c65be06c428a"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.003160 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6918e56a-00f4-4f99-b3ef-c65be06c428a" (UID: "6918e56a-00f4-4f99-b3ef-c65be06c428a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.003705 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-kube-api-access-nxs7n" (OuterVolumeSpecName: "kube-api-access-nxs7n") pod "6918e56a-00f4-4f99-b3ef-c65be06c428a" (UID: "6918e56a-00f4-4f99-b3ef-c65be06c428a"). InnerVolumeSpecName "kube-api-access-nxs7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.099878 4944 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6918e56a-00f4-4f99-b3ef-c65be06c428a-cache\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.099914 4944 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.099938 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.099948 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxs7n\" (UniqueName: \"kubernetes.io/projected/6918e56a-00f4-4f99-b3ef-c65be06c428a-kube-api-access-nxs7n\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.126713 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.170292 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d5d955fb7-kf4q4_69a7c779-b5d5-4d2e-85a5-bd90df91a4c6/neutron-api/0.log" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.170366 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.201256 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.302575 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-combined-ca-bundle\") pod \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.302651 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-config\") pod \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.302758 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-httpd-config\") pod \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.302795 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-ovndb-tls-certs\") pod \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.302852 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st78n\" (UniqueName: \"kubernetes.io/projected/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-kube-api-access-st78n\") pod \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.302925 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-internal-tls-certs\") pod \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.302967 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-public-tls-certs\") pod \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\" (UID: \"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6\") " Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.305832 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" (UID: "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.317288 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-kube-api-access-st78n" (OuterVolumeSpecName: "kube-api-access-st78n") pod "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" (UID: "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6"). InnerVolumeSpecName "kube-api-access-st78n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.342551 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" (UID: "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.343060 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-config" (OuterVolumeSpecName: "config") pod "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" (UID: "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.343889 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" (UID: "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.344230 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" (UID: "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.360376 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" (UID: "69a7c779-b5d5-4d2e-85a5-bd90df91a4c6"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.404923 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.404963 4944 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.404975 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st78n\" (UniqueName: \"kubernetes.io/projected/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-kube-api-access-st78n\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.404985 4944 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.404993 4944 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.405003 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.405012 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6-config\") on node \"crc\" DevicePath \"\"" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.549206 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.549273 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.549329 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.550142 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f8ac17962023518b7abc2c1aa6e647faf7d9a5693323d499415d0670598b6a65"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.550206 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://f8ac17962023518b7abc2c1aa6e647faf7d9a5693323d499415d0670598b6a65" gracePeriod=600 Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.691236 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d5d955fb7-kf4q4_69a7c779-b5d5-4d2e-85a5-bd90df91a4c6/neutron-api/0.log" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.691521 4944 generic.go:334] "Generic (PLEG): container finished" podID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" containerID="ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d" exitCode=137 Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.691567 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d5d955fb7-kf4q4" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.691590 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d5d955fb7-kf4q4" event={"ID":"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6","Type":"ContainerDied","Data":"ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d"} Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.691904 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d5d955fb7-kf4q4" event={"ID":"69a7c779-b5d5-4d2e-85a5-bd90df91a4c6","Type":"ContainerDied","Data":"5b0f096028841e609c6de90a324e48f2efd19aa29e3d69de72b1f9c68f012de8"} Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.691936 4944 scope.go:117] "RemoveContainer" containerID="dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.697935 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="f8ac17962023518b7abc2c1aa6e647faf7d9a5693323d499415d0670598b6a65" exitCode=0 Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.697974 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"f8ac17962023518b7abc2c1aa6e647faf7d9a5693323d499415d0670598b6a65"} Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.705755 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6918e56a-00f4-4f99-b3ef-c65be06c428a","Type":"ContainerDied","Data":"f531019f16c6b75dbbdfa316451f4b78f6ee7377c8c252e084ad48887aa300de"} Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.705877 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.730024 4944 scope.go:117] "RemoveContainer" containerID="ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.730240 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d5d955fb7-kf4q4"] Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.738672 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5d5d955fb7-kf4q4"] Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.756678 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.760436 4944 scope.go:117] "RemoveContainer" containerID="dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992" Nov 24 09:14:53 crc kubenswrapper[4944]: E1124 09:14:53.762384 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992\": container with ID starting with dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992 not found: ID does not exist" containerID="dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.762437 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992"} err="failed to get container status \"dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992\": rpc error: code = NotFound desc = could not find container \"dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992\": container with ID starting with dbfeda7e74aa15c985ad3d564cbb836a8e49bb56118edfa0e4d8b2d6478a4992 not found: ID does not exist" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.762469 4944 scope.go:117] "RemoveContainer" containerID="ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d" Nov 24 09:14:53 crc kubenswrapper[4944]: E1124 09:14:53.762950 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d\": container with ID starting with ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d not found: ID does not exist" containerID="ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.762978 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d"} err="failed to get container status \"ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d\": rpc error: code = NotFound desc = could not find container \"ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d\": container with ID starting with ba6f2dc59dd0b22091368b6de9a539f7a98eb834f29abc2be5c2e0007bd5134d not found: ID does not exist" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.762994 4944 scope.go:117] "RemoveContainer" containerID="1d4676a5ed72cda8e3ecd73f73356ca831b52b82f176ed2715695ec66c7c950e" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.765204 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.794159 4944 scope.go:117] "RemoveContainer" containerID="d6d5476d01856b9678b39efa998de2c30d0cc6772d172d0bf1aa500eb8ebaca7" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.816068 4944 scope.go:117] "RemoveContainer" containerID="be411b5a8274b541c609635739f60faee6efa6abf3fbf8c48eed07bef8944385" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.833748 4944 scope.go:117] "RemoveContainer" containerID="48d0fed6b1c1123b8f3c416f72272d6749327411fdfbf987f7ee43dd423e071e" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.858731 4944 scope.go:117] "RemoveContainer" containerID="6ccdc9b46a378806550b2c6500f4fd45536785660f1ba53f9c389a60f6a9bf41" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.876297 4944 scope.go:117] "RemoveContainer" containerID="15e66884ade4e39577226708f15ca7622732f6b82d81fb79dff82cc2578f9a9a" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.895439 4944 scope.go:117] "RemoveContainer" containerID="e6c6e71ef8c3c8177a39c99dabb96e105d42c681788bccbca30c6f9504cba707" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.912190 4944 scope.go:117] "RemoveContainer" containerID="bb4e495759495e1efa8e3c7bf1b539d36a09c8a14b177fd703baec61783412a9" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.929830 4944 scope.go:117] "RemoveContainer" containerID="c17e6f2fe0c81b82e5511368c8268b7225365160cc97761f7dc0ee157f7a8484" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.945638 4944 scope.go:117] "RemoveContainer" containerID="91b30be5efcaf3b6adec37e7c1e859fed2c520997d002c00b9500d39d521b0ed" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.964388 4944 scope.go:117] "RemoveContainer" containerID="e533818c9ecf15f88341998771d0639c9cac4cfb451d5374ad40bec2e815a3ff" Nov 24 09:14:53 crc kubenswrapper[4944]: I1124 09:14:53.983618 4944 scope.go:117] "RemoveContainer" containerID="5d2fbee82caba7d0ca08ca333bcbf3e804351883721f48d8d9bfaac47a78f960" Nov 24 09:14:54 crc kubenswrapper[4944]: I1124 09:14:54.001870 4944 scope.go:117] "RemoveContainer" containerID="87e895d554127f520ab155e47e47d2faa5077bd4196802ecc68ee9b04403fe3d" Nov 24 09:14:54 crc kubenswrapper[4944]: I1124 09:14:54.022093 4944 scope.go:117] "RemoveContainer" containerID="9311cf4164c78aa256b1228bc545098e231c1036813339b9399c8c0ccdf8463f" Nov 24 09:14:54 crc kubenswrapper[4944]: I1124 09:14:54.047533 4944 scope.go:117] "RemoveContainer" containerID="2842b8bb44b04f4fb4de40704e3d22283e14c59ce261e4a211bef21dbb6e7a96" Nov 24 09:14:54 crc kubenswrapper[4944]: I1124 09:14:54.069352 4944 scope.go:117] "RemoveContainer" containerID="3d31ceee24450aa9fe538636500f7868bbd37f5da8e0012f05ca3b457f8720eb" Nov 24 09:14:54 crc kubenswrapper[4944]: I1124 09:14:54.295270 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53d02d15-b171-4e78-b528-9bec6362bf70" path="/var/lib/kubelet/pods/53d02d15-b171-4e78-b528-9bec6362bf70/volumes" Nov 24 09:14:54 crc kubenswrapper[4944]: I1124 09:14:54.296694 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" path="/var/lib/kubelet/pods/6918e56a-00f4-4f99-b3ef-c65be06c428a/volumes" Nov 24 09:14:54 crc kubenswrapper[4944]: I1124 09:14:54.299514 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" path="/var/lib/kubelet/pods/69a7c779-b5d5-4d2e-85a5-bd90df91a4c6/volumes" Nov 24 09:14:54 crc kubenswrapper[4944]: I1124 09:14:54.300009 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" path="/var/lib/kubelet/pods/e1d0474c-20bb-4653-8ce7-176bf4bc5f0b/volumes" Nov 24 09:14:54 crc kubenswrapper[4944]: I1124 09:14:54.716996 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731"} Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.139256 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr"] Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140149 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18d26ea6-eff1-4ad5-b50b-dd233a22e982" containerName="galera" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140164 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="18d26ea6-eff1-4ad5-b50b-dd233a22e982" containerName="galera" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140174 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140181 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140205 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9d45d4c-ab9f-426c-a193-7f46c398ed64" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140213 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9d45d4c-ab9f-426c-a193-7f46c398ed64" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140221 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c04aca3-abaa-4c62-8e62-af920276cc50" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140227 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c04aca3-abaa-4c62-8e62-af920276cc50" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140236 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-replicator" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140242 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-replicator" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140254 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerName="proxy-server" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140260 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerName="proxy-server" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140268 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ea29cb-ae71-48af-8005-ba99c178a573" containerName="barbican-keystone-listener-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140274 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ea29cb-ae71-48af-8005-ba99c178a573" containerName="barbican-keystone-listener-log" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140280 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-updater" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140286 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-updater" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140296 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e454a738-0872-41d8-9432-17432276248c" containerName="galera" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140302 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e454a738-0872-41d8-9432-17432276248c" containerName="galera" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140314 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="proxy-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140319 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="proxy-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140329 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-replicator" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140334 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-replicator" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140342 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerName="barbican-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140384 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerName="barbican-api" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140395 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90fb3cc2-0725-48dc-9ad5-5c8da27930f8" containerName="nova-scheduler-scheduler" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140402 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="90fb3cc2-0725-48dc-9ad5-5c8da27930f8" containerName="nova-scheduler-scheduler" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140410 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerName="nova-api-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140416 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerName="nova-api-api" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140423 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server-init" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140428 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server-init" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140438 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="ceilometer-central-agent" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140446 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="ceilometer-central-agent" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140454 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-metadata" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140459 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-metadata" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140469 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a65aeee-1ac4-4ae3-b713-ad1d0074f797" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140475 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a65aeee-1ac4-4ae3-b713-ad1d0074f797" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140483 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" containerName="dnsmasq-dns" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140489 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" containerName="dnsmasq-dns" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140495 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e92d3f-3532-467f-bf38-c9c3107d4fcb" containerName="rabbitmq" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140500 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e92d3f-3532-467f-bf38-c9c3107d4fcb" containerName="rabbitmq" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140506 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-server" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140512 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-server" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140520 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-auditor" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140526 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-auditor" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140535 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32" containerName="nova-cell1-conductor-conductor" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140540 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32" containerName="nova-cell1-conductor-conductor" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140549 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-server" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140554 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-server" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140563 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" containerName="neutron-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140570 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" containerName="neutron-api" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140585 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerName="nova-api-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140593 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerName="nova-api-log" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140606 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35137b92-4f56-4399-8121-89bdb126162b" containerName="memcached" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140613 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="35137b92-4f56-4399-8121-89bdb126162b" containerName="memcached" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140625 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerName="proxy-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140631 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerName="proxy-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140638 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-server" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140676 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-server" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140688 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140694 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-log" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140704 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2495f8a-6f01-47e6-beff-f0d7345967a7" containerName="keystone-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140710 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2495f8a-6f01-47e6-beff-f0d7345967a7" containerName="keystone-api" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140722 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c521e9a3-0d64-4658-8a00-690d20e619d0" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140729 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c521e9a3-0d64-4658-8a00-690d20e619d0" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140742 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="rsync" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140749 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="rsync" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140769 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerName="barbican-api-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140774 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerName="barbican-api-log" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140784 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d71630-681e-4126-bb64-e8b03f83ce97" containerName="nova-cell0-conductor-conductor" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140792 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d71630-681e-4126-bb64-e8b03f83ce97" containerName="nova-cell0-conductor-conductor" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140804 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f94b9960-fa37-49f7-9fa7-0e01f855992b" containerName="glance-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140812 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f94b9960-fa37-49f7-9fa7-0e01f855992b" containerName="glance-log" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140820 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f94b9960-fa37-49f7-9fa7-0e01f855992b" containerName="glance-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140826 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f94b9960-fa37-49f7-9fa7-0e01f855992b" containerName="glance-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140836 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11d8c8f9-ee83-4450-9a76-3c0a020652de" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140842 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="11d8c8f9-ee83-4450-9a76-3c0a020652de" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140851 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9d45d4c-ab9f-426c-a193-7f46c398ed64" containerName="ovsdbserver-nb" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140858 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9d45d4c-ab9f-426c-a193-7f46c398ed64" containerName="ovsdbserver-nb" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140873 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9c665f-2be7-4b08-934d-b63036656b5f" containerName="barbican-worker" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140880 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9c665f-2be7-4b08-934d-b63036656b5f" containerName="barbican-worker" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140889 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a8d972-b16f-421e-9fe4-f57653048845" containerName="cinder-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140895 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a8d972-b16f-421e-9fe4-f57653048845" containerName="cinder-api" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140903 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7209b8a6-5a97-4611-8183-84f1db25e2d1" containerName="glance-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140911 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7209b8a6-5a97-4611-8183-84f1db25e2d1" containerName="glance-log" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140924 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-auditor" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140931 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-auditor" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140939 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b2d40c3-9522-46f4-85e7-fd724cb1bdf2" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140946 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b2d40c3-9522-46f4-85e7-fd724cb1bdf2" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140965 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e0b8db-ac91-432d-a87c-e741e441c33a" containerName="kube-state-metrics" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140977 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e0b8db-ac91-432d-a87c-e741e441c33a" containerName="kube-state-metrics" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.140990 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a8d972-b16f-421e-9fe4-f57653048845" containerName="cinder-api-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.140998 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a8d972-b16f-421e-9fe4-f57653048845" containerName="cinder-api-log" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141013 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d02d15-b171-4e78-b528-9bec6362bf70" containerName="cinder-scheduler" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141021 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d02d15-b171-4e78-b528-9bec6362bf70" containerName="cinder-scheduler" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141030 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-replicator" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141039 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-replicator" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141070 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="swift-recon-cron" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141079 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="swift-recon-cron" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141087 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ea29cb-ae71-48af-8005-ba99c178a573" containerName="barbican-keystone-listener" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141095 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ea29cb-ae71-48af-8005-ba99c178a573" containerName="barbican-keystone-listener" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141104 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69996228-3cfa-486e-8291-12e44f438c5b" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141112 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="69996228-3cfa-486e-8291-12e44f438c5b" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141123 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e92d3f-3532-467f-bf38-c9c3107d4fcb" containerName="setup-container" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141129 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e92d3f-3532-467f-bf38-c9c3107d4fcb" containerName="setup-container" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141139 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovs-vswitchd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141145 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovs-vswitchd" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141155 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-reaper" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141160 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-reaper" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141173 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="ceilometer-notification-agent" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141182 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="ceilometer-notification-agent" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141191 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e454a738-0872-41d8-9432-17432276248c" containerName="mysql-bootstrap" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141198 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e454a738-0872-41d8-9432-17432276248c" containerName="mysql-bootstrap" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141205 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-auditor" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141211 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-auditor" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141244 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" containerName="init" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141251 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" containerName="init" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141262 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-expirer" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141301 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-expirer" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141311 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a59d46-95bf-4681-b61d-22239c638737" containerName="placement-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141317 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a59d46-95bf-4681-b61d-22239c638737" containerName="placement-log" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141331 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141338 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141346 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18d26ea6-eff1-4ad5-b50b-dd233a22e982" containerName="mysql-bootstrap" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141352 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="18d26ea6-eff1-4ad5-b50b-dd233a22e982" containerName="mysql-bootstrap" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141360 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b69c572d-53a7-4d39-a8fc-4152846915a5" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141367 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b69c572d-53a7-4d39-a8fc-4152846915a5" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141381 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a59d46-95bf-4681-b61d-22239c638737" containerName="placement-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141389 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a59d46-95bf-4681-b61d-22239c638737" containerName="placement-api" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141399 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d02d15-b171-4e78-b528-9bec6362bf70" containerName="probe" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141405 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d02d15-b171-4e78-b528-9bec6362bf70" containerName="probe" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141414 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709b6d2b-7e0f-4f60-9bd2-edacb409a727" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141420 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="709b6d2b-7e0f-4f60-9bd2-edacb409a727" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141427 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="sg-core" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141433 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="sg-core" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141440 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c04aca3-abaa-4c62-8e62-af920276cc50" containerName="ovsdbserver-sb" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141446 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c04aca3-abaa-4c62-8e62-af920276cc50" containerName="ovsdbserver-sb" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141452 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7209b8a6-5a97-4611-8183-84f1db25e2d1" containerName="glance-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141458 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7209b8a6-5a97-4611-8183-84f1db25e2d1" containerName="glance-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141464 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-updater" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141470 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-updater" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141479 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" containerName="neutron-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141485 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" containerName="neutron-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141492 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9c665f-2be7-4b08-934d-b63036656b5f" containerName="barbican-worker-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141497 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9c665f-2be7-4b08-934d-b63036656b5f" containerName="barbican-worker-log" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141506 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerName="ovn-northd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141512 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerName="ovn-northd" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141525 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69084f7b-b347-4fdf-917e-2f534f3cc47c" containerName="ovn-controller" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141531 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="69084f7b-b347-4fdf-917e-2f534f3cc47c" containerName="ovn-controller" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141543 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a9098c-7257-46f7-888c-c4bb6f9421a5" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141549 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a9098c-7257-46f7-888c-c4bb6f9421a5" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: E1124 09:15:00.141558 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1db23d75-43f2-4bae-9e09-399d0ffb7ed5" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141568 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1db23d75-43f2-4bae-9e09-399d0ffb7ed5" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141734 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d9c665f-2be7-4b08-934d-b63036656b5f" containerName="barbican-worker" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141747 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerName="nova-api-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141757 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="7209b8a6-5a97-4611-8183-84f1db25e2d1" containerName="glance-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141767 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="69084f7b-b347-4fdf-917e-2f534f3cc47c" containerName="ovn-controller" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141776 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovs-vswitchd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141784 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-server" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141791 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d9c665f-2be7-4b08-934d-b63036656b5f" containerName="barbican-worker-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141799 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f12c5bf-4d28-4c6d-bea1-d9ce4cc992ef" containerName="dnsmasq-dns" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141807 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="69996228-3cfa-486e-8291-12e44f438c5b" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141818 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-server" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141825 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce23facc-1a0d-4c53-b696-b37a7525eba7" containerName="nova-api-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141836 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e454a738-0872-41d8-9432-17432276248c" containerName="galera" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141847 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9d45d4c-ab9f-426c-a193-7f46c398ed64" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141857 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="rsync" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141870 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="ceilometer-central-agent" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141879 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c04aca3-abaa-4c62-8e62-af920276cc50" containerName="ovsdbserver-sb" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141888 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f94b9960-fa37-49f7-9fa7-0e01f855992b" containerName="glance-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141896 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="98fd9f3f-ba8e-47b2-b375-cf1dd9c2ef32" containerName="nova-cell1-conductor-conductor" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141905 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f94b9960-fa37-49f7-9fa7-0e01f855992b" containerName="glance-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141916 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-auditor" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141925 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a8d972-b16f-421e-9fe4-f57653048845" containerName="cinder-api-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141932 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerName="proxy-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141938 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="70a59d46-95bf-4681-b61d-22239c638737" containerName="placement-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141945 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-auditor" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141953 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" containerName="neutron-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141960 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2495f8a-6f01-47e6-beff-f0d7345967a7" containerName="keystone-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141967 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141973 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="sg-core" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141982 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d0474c-20bb-4653-8ce7-176bf4bc5f0b" containerName="ovsdb-server" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141990 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerName="barbican-api-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.141997 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b69c572d-53a7-4d39-a8fc-4152846915a5" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142005 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-metadata" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142013 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-replicator" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142022 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-expirer" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142029 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="90fb3cc2-0725-48dc-9ad5-5c8da27930f8" containerName="nova-scheduler-scheduler" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142035 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="proxy-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142059 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1db23d75-43f2-4bae-9e09-399d0ffb7ed5" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142067 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="55a9098c-7257-46f7-888c-c4bb6f9421a5" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142075 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="7209b8a6-5a97-4611-8183-84f1db25e2d1" containerName="glance-httpd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142085 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="70a59d46-95bf-4681-b61d-22239c638737" containerName="placement-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142091 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d71630-681e-4126-bb64-e8b03f83ce97" containerName="nova-cell0-conductor-conductor" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142101 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="account-reaper" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142109 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="18d26ea6-eff1-4ad5-b50b-dd233a22e982" containerName="galera" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142118 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dfd2e16-b145-4464-8c5a-bff9a5908d45" containerName="ovn-northd" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142125 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-auditor" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142136 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="046a8c15-6acd-47fe-aef7-daac5085c2ec" containerName="barbican-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142145 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d1b549-198c-4a0d-a895-2b1f10dc337d" containerName="proxy-server" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142151 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-server" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142161 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-replicator" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142170 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ea29cb-ae71-48af-8005-ba99c178a573" containerName="barbican-keystone-listener" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142180 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="11d8c8f9-ee83-4450-9a76-3c0a020652de" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142191 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e92d3f-3532-467f-bf38-c9c3107d4fcb" containerName="rabbitmq" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142200 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-updater" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142208 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e0b8db-ac91-432d-a87c-e741e441c33a" containerName="kube-state-metrics" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142215 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb23d74-ac96-4ce2-8a3b-0fe9aea10fe4" containerName="ceilometer-notification-agent" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142224 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="container-replicator" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142232 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="709b6d2b-7e0f-4f60-9bd2-edacb409a727" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142241 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9d45d4c-ab9f-426c-a193-7f46c398ed64" containerName="ovsdbserver-nb" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142248 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="swift-recon-cron" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142255 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="35137b92-4f56-4399-8121-89bdb126162b" containerName="memcached" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142263 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6918e56a-00f4-4f99-b3ef-c65be06c428a" containerName="object-updater" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142272 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a65aeee-1ac4-4ae3-b713-ad1d0074f797" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142283 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c04aca3-abaa-4c62-8e62-af920276cc50" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142293 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="c521e9a3-0d64-4658-8a00-690d20e619d0" containerName="openstack-network-exporter" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142302 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ea29cb-ae71-48af-8005-ba99c178a573" containerName="barbican-keystone-listener-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142310 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d02d15-b171-4e78-b528-9bec6362bf70" containerName="cinder-scheduler" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142317 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b2d40c3-9522-46f4-85e7-fd724cb1bdf2" containerName="mariadb-account-delete" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142322 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc12ffc5-e026-42a6-8ddc-43961c535787" containerName="nova-metadata-log" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142365 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d02d15-b171-4e78-b528-9bec6362bf70" containerName="probe" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142371 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="69a7c779-b5d5-4d2e-85a5-bd90df91a4c6" containerName="neutron-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142381 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a8d972-b16f-421e-9fe4-f57653048845" containerName="cinder-api" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.142927 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.145137 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.145260 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.152215 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr"] Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.196630 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dxqz\" (UniqueName: \"kubernetes.io/projected/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-kube-api-access-9dxqz\") pod \"collect-profiles-29399595-gqrtr\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.196742 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-config-volume\") pod \"collect-profiles-29399595-gqrtr\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.196835 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-secret-volume\") pod \"collect-profiles-29399595-gqrtr\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.298522 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-config-volume\") pod \"collect-profiles-29399595-gqrtr\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.298612 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-secret-volume\") pod \"collect-profiles-29399595-gqrtr\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.298693 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dxqz\" (UniqueName: \"kubernetes.io/projected/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-kube-api-access-9dxqz\") pod \"collect-profiles-29399595-gqrtr\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.299925 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-config-volume\") pod \"collect-profiles-29399595-gqrtr\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.308165 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-secret-volume\") pod \"collect-profiles-29399595-gqrtr\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.322195 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dxqz\" (UniqueName: \"kubernetes.io/projected/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-kube-api-access-9dxqz\") pod \"collect-profiles-29399595-gqrtr\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.464015 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:00 crc kubenswrapper[4944]: I1124 09:15:00.897625 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr"] Nov 24 09:15:01 crc kubenswrapper[4944]: I1124 09:15:01.789106 4944 generic.go:334] "Generic (PLEG): container finished" podID="b2319a23-bfcf-45bc-889d-3c2f1bc8c582" containerID="a2861eff1e5debc3898f43ae95d6265c65e00df89ffe968ea6df630b1443c7c0" exitCode=0 Nov 24 09:15:01 crc kubenswrapper[4944]: I1124 09:15:01.789216 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" event={"ID":"b2319a23-bfcf-45bc-889d-3c2f1bc8c582","Type":"ContainerDied","Data":"a2861eff1e5debc3898f43ae95d6265c65e00df89ffe968ea6df630b1443c7c0"} Nov 24 09:15:01 crc kubenswrapper[4944]: I1124 09:15:01.790509 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" event={"ID":"b2319a23-bfcf-45bc-889d-3c2f1bc8c582","Type":"ContainerStarted","Data":"04a88a247a81138ba3969cea2bfdba623f15af952bd6676c652698724c4b5503"} Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.089314 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.138170 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-secret-volume\") pod \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.138218 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-config-volume\") pod \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.139040 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-config-volume" (OuterVolumeSpecName: "config-volume") pod "b2319a23-bfcf-45bc-889d-3c2f1bc8c582" (UID: "b2319a23-bfcf-45bc-889d-3c2f1bc8c582"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.139187 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dxqz\" (UniqueName: \"kubernetes.io/projected/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-kube-api-access-9dxqz\") pod \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\" (UID: \"b2319a23-bfcf-45bc-889d-3c2f1bc8c582\") " Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.139494 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.144221 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-kube-api-access-9dxqz" (OuterVolumeSpecName: "kube-api-access-9dxqz") pod "b2319a23-bfcf-45bc-889d-3c2f1bc8c582" (UID: "b2319a23-bfcf-45bc-889d-3c2f1bc8c582"). InnerVolumeSpecName "kube-api-access-9dxqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.144382 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b2319a23-bfcf-45bc-889d-3c2f1bc8c582" (UID: "b2319a23-bfcf-45bc-889d-3c2f1bc8c582"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.241409 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.241451 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dxqz\" (UniqueName: \"kubernetes.io/projected/b2319a23-bfcf-45bc-889d-3c2f1bc8c582-kube-api-access-9dxqz\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.811301 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" event={"ID":"b2319a23-bfcf-45bc-889d-3c2f1bc8c582","Type":"ContainerDied","Data":"04a88a247a81138ba3969cea2bfdba623f15af952bd6676c652698724c4b5503"} Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.811362 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04a88a247a81138ba3969cea2bfdba623f15af952bd6676c652698724c4b5503" Nov 24 09:15:03 crc kubenswrapper[4944]: I1124 09:15:03.811400 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr" Nov 24 09:15:25 crc kubenswrapper[4944]: E1124 09:15:25.081498 4944 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 24 09:15:25 crc kubenswrapper[4944]: E1124 09:15:25.082070 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data podName:12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5 nodeName:}" failed. No retries permitted until 2025-11-24 09:16:29.082038345 +0000 UTC m=+1449.616478807 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data") pod "rabbitmq-cell1-server-0" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5") : configmap "rabbitmq-cell1-config-data" not found Nov 24 09:15:26 crc kubenswrapper[4944]: E1124 09:15:26.131087 4944 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 24 09:15:26 crc kubenswrapper[4944]: command '/bin/bash -c if [ ! -z "$(cat /etc/pod-info/skipPreStopChecks)" ]; then exit 0; fi; rabbitmq-upgrade await_online_quorum_plus_one -t 604800 && rabbitmq-upgrade await_online_synchronized_mirror -t 604800 || true && rabbitmq-upgrade drain -t 604800' exited with 69: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Most common reasons for this are: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 24 09:15:26 crc kubenswrapper[4944]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is not running Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: In addition to the diagnostics info below: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 24 09:15:26 crc kubenswrapper[4944]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 24 09:15:26 crc kubenswrapper[4944]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: DIAGNOSTICS Nov 24 09:15:26 crc kubenswrapper[4944]: =========== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 24 09:15:26 crc kubenswrapper[4944]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Current node details: Nov 24 09:15:26 crc kubenswrapper[4944]: * node name: 'rabbitmqcli-200-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 24 09:15:26 crc kubenswrapper[4944]: * effective user's home directory: /var/lib/rabbitmq Nov 24 09:15:26 crc kubenswrapper[4944]: * Erlang cookie hash: KwOdyRGKh/mgrdrrr9UoWA== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Most common reasons for this are: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 24 09:15:26 crc kubenswrapper[4944]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is not running Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: In addition to the diagnostics info below: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 24 09:15:26 crc kubenswrapper[4944]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 24 09:15:26 crc kubenswrapper[4944]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: DIAGNOSTICS Nov 24 09:15:26 crc kubenswrapper[4944]: =========== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 24 09:15:26 crc kubenswrapper[4944]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Current node details: Nov 24 09:15:26 crc kubenswrapper[4944]: * node name: 'rabbitmqcli-113-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 24 09:15:26 crc kubenswrapper[4944]: * effective user's home directory: /var/lib/rabbitmq Nov 24 09:15:26 crc kubenswrapper[4944]: * Erlang cookie hash: KwOdyRGKh/mgrdrrr9UoWA== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: > execCommand=["/bin/bash","-c","if [ ! -z \"$(cat /etc/pod-info/skipPreStopChecks)\" ]; then exit 0; fi; rabbitmq-upgrade await_online_quorum_plus_one -t 604800 \u0026\u0026 rabbitmq-upgrade await_online_synchronized_mirror -t 604800 || true \u0026\u0026 rabbitmq-upgrade drain -t 604800"] containerName="rabbitmq" pod="openstack/rabbitmq-cell1-server-0" message=< Nov 24 09:15:26 crc kubenswrapper[4944]: Will put node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack into maintenance mode. The node will no longer serve any client traffic! Nov 24 09:15:26 crc kubenswrapper[4944]: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Most common reasons for this are: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 24 09:15:26 crc kubenswrapper[4944]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is not running Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: In addition to the diagnostics info below: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 24 09:15:26 crc kubenswrapper[4944]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 24 09:15:26 crc kubenswrapper[4944]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: DIAGNOSTICS Nov 24 09:15:26 crc kubenswrapper[4944]: =========== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 24 09:15:26 crc kubenswrapper[4944]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Current node details: Nov 24 09:15:26 crc kubenswrapper[4944]: * node name: 'rabbitmqcli-200-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 24 09:15:26 crc kubenswrapper[4944]: * effective user's home directory: /var/lib/rabbitmq Nov 24 09:15:26 crc kubenswrapper[4944]: * Erlang cookie hash: KwOdyRGKh/mgrdrrr9UoWA== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Most common reasons for this are: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 24 09:15:26 crc kubenswrapper[4944]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is not running Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: In addition to the diagnostics info below: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 24 09:15:26 crc kubenswrapper[4944]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 24 09:15:26 crc kubenswrapper[4944]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: DIAGNOSTICS Nov 24 09:15:26 crc kubenswrapper[4944]: =========== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 24 09:15:26 crc kubenswrapper[4944]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Current node details: Nov 24 09:15:26 crc kubenswrapper[4944]: * node name: 'rabbitmqcli-113-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 24 09:15:26 crc kubenswrapper[4944]: * effective user's home directory: /var/lib/rabbitmq Nov 24 09:15:26 crc kubenswrapper[4944]: * Erlang cookie hash: KwOdyRGKh/mgrdrrr9UoWA== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: > Nov 24 09:15:26 crc kubenswrapper[4944]: E1124 09:15:26.131135 4944 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 24 09:15:26 crc kubenswrapper[4944]: command '/bin/bash -c if [ ! -z "$(cat /etc/pod-info/skipPreStopChecks)" ]; then exit 0; fi; rabbitmq-upgrade await_online_quorum_plus_one -t 604800 && rabbitmq-upgrade await_online_synchronized_mirror -t 604800 || true && rabbitmq-upgrade drain -t 604800' exited with 69: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Most common reasons for this are: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 24 09:15:26 crc kubenswrapper[4944]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is not running Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: In addition to the diagnostics info below: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 24 09:15:26 crc kubenswrapper[4944]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 24 09:15:26 crc kubenswrapper[4944]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: DIAGNOSTICS Nov 24 09:15:26 crc kubenswrapper[4944]: =========== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 24 09:15:26 crc kubenswrapper[4944]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Current node details: Nov 24 09:15:26 crc kubenswrapper[4944]: * node name: 'rabbitmqcli-200-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 24 09:15:26 crc kubenswrapper[4944]: * effective user's home directory: /var/lib/rabbitmq Nov 24 09:15:26 crc kubenswrapper[4944]: * Erlang cookie hash: KwOdyRGKh/mgrdrrr9UoWA== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Error: unable to perform an operation on node 'rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'. Please see diagnostics information and suggestions below. Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Most common reasons for this are: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is unreachable (e.g. due to hostname resolution, TCP connection or firewall issues) Nov 24 09:15:26 crc kubenswrapper[4944]: * CLI tool fails to authenticate with the server (e.g. due to CLI tool's Erlang cookie not matching that of the server) Nov 24 09:15:26 crc kubenswrapper[4944]: * Target node is not running Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: In addition to the diagnostics info below: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: * See the CLI, clustering and networking guides on https://rabbitmq.com/documentation.html to learn more Nov 24 09:15:26 crc kubenswrapper[4944]: * Consult server logs on node rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack Nov 24 09:15:26 crc kubenswrapper[4944]: * If target node is configured to use long node names, don't forget to use --longnames with CLI tools Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: DIAGNOSTICS Nov 24 09:15:26 crc kubenswrapper[4944]: =========== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: attempted to contact: ['rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack'] Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: Nov 24 09:15:26 crc kubenswrapper[4944]: * unable to connect to epmd (port 4369) on rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack: nxdomain (non-existing domain) Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: Current node details: Nov 24 09:15:26 crc kubenswrapper[4944]: * node name: 'rabbitmqcli-113-rabbit@rabbitmq-cell1-server-0.rabbitmq-cell1-nodes.openstack' Nov 24 09:15:26 crc kubenswrapper[4944]: * effective user's home directory: /var/lib/rabbitmq Nov 24 09:15:26 crc kubenswrapper[4944]: * Erlang cookie hash: KwOdyRGKh/mgrdrrr9UoWA== Nov 24 09:15:26 crc kubenswrapper[4944]: Nov 24 09:15:26 crc kubenswrapper[4944]: > pod="openstack/rabbitmq-cell1-server-0" podUID="12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" containerName="rabbitmq" containerID="cri-o://07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256" Nov 24 09:15:26 crc kubenswrapper[4944]: I1124 09:15:26.131173 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" containerName="rabbitmq" containerID="cri-o://07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256" gracePeriod=604738 Nov 24 09:15:29 crc kubenswrapper[4944]: I1124 09:15:29.152024 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.644364 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.778980 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.779460 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-confd\") pod \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.779492 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-tls\") pod \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.779523 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-erlang-cookie-secret\") pod \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.779564 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-plugins-conf\") pod \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.779609 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-server-conf\") pod \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.779648 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-plugins\") pod \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.779675 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data\") pod \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.779700 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-pod-info\") pod \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.779716 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgfx2\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-kube-api-access-jgfx2\") pod \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.779792 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-erlang-cookie\") pod \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\" (UID: \"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5\") " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.780645 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.780805 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.781149 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.785605 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.786177 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-pod-info" (OuterVolumeSpecName: "pod-info") pod "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.786193 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.786566 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.787040 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-kube-api-access-jgfx2" (OuterVolumeSpecName: "kube-api-access-jgfx2") pod "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5"). InnerVolumeSpecName "kube-api-access-jgfx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.808722 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data" (OuterVolumeSpecName: "config-data") pod "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.826851 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-server-conf" (OuterVolumeSpecName: "server-conf") pod "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.853782 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" (UID: "12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.881982 4944 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.882023 4944 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.882036 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.882083 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.882095 4944 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.882106 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgfx2\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-kube-api-access-jgfx2\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.882123 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.882164 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.882178 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.882190 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.882201 4944 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.900146 4944 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 24 09:15:32 crc kubenswrapper[4944]: I1124 09:15:32.983585 4944 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.111506 4944 generic.go:334] "Generic (PLEG): container finished" podID="12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" containerID="07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256" exitCode=0 Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.111546 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5","Type":"ContainerDied","Data":"07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256"} Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.111586 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.111597 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5","Type":"ContainerDied","Data":"47d0fc0355d2949def167fe23b4bb446e7348b2e8b764656ab97634732468cbc"} Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.111618 4944 scope.go:117] "RemoveContainer" containerID="07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256" Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.132837 4944 scope.go:117] "RemoveContainer" containerID="0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033" Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.148026 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.155532 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.171753 4944 scope.go:117] "RemoveContainer" containerID="07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256" Nov 24 09:15:33 crc kubenswrapper[4944]: E1124 09:15:33.172248 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256\": container with ID starting with 07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256 not found: ID does not exist" containerID="07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256" Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.172297 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256"} err="failed to get container status \"07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256\": rpc error: code = NotFound desc = could not find container \"07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256\": container with ID starting with 07c1d45cde236f0ef8250c8425c298b1f30450984f34d9f7ae41c2f2f1e7b256 not found: ID does not exist" Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.172322 4944 scope.go:117] "RemoveContainer" containerID="0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033" Nov 24 09:15:33 crc kubenswrapper[4944]: E1124 09:15:33.172623 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033\": container with ID starting with 0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033 not found: ID does not exist" containerID="0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033" Nov 24 09:15:33 crc kubenswrapper[4944]: I1124 09:15:33.172661 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033"} err="failed to get container status \"0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033\": rpc error: code = NotFound desc = could not find container \"0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033\": container with ID starting with 0243e5b384f1105c91e95aa830531d67feeb721ddc9c4b01da219c93f325c033 not found: ID does not exist" Nov 24 09:15:34 crc kubenswrapper[4944]: I1124 09:15:34.286002 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" path="/var/lib/kubelet/pods/12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5/volumes" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.612199 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rfsn8"] Nov 24 09:15:42 crc kubenswrapper[4944]: E1124 09:15:42.613112 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" containerName="setup-container" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.613131 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" containerName="setup-container" Nov 24 09:15:42 crc kubenswrapper[4944]: E1124 09:15:42.613162 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2319a23-bfcf-45bc-889d-3c2f1bc8c582" containerName="collect-profiles" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.613171 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2319a23-bfcf-45bc-889d-3c2f1bc8c582" containerName="collect-profiles" Nov 24 09:15:42 crc kubenswrapper[4944]: E1124 09:15:42.613191 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" containerName="rabbitmq" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.613199 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" containerName="rabbitmq" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.613358 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="12fa9f2b-3cb2-4f20-827b-961ee1f9c7c5" containerName="rabbitmq" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.613386 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2319a23-bfcf-45bc-889d-3c2f1bc8c582" containerName="collect-profiles" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.615235 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.625246 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rfsn8"] Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.737881 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c4wq\" (UniqueName: \"kubernetes.io/projected/2b33ed49-adf4-4398-8723-e53ef109cc70-kube-api-access-4c4wq\") pod \"redhat-operators-rfsn8\" (UID: \"2b33ed49-adf4-4398-8723-e53ef109cc70\") " pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.738103 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b33ed49-adf4-4398-8723-e53ef109cc70-catalog-content\") pod \"redhat-operators-rfsn8\" (UID: \"2b33ed49-adf4-4398-8723-e53ef109cc70\") " pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.738136 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b33ed49-adf4-4398-8723-e53ef109cc70-utilities\") pod \"redhat-operators-rfsn8\" (UID: \"2b33ed49-adf4-4398-8723-e53ef109cc70\") " pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.839273 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b33ed49-adf4-4398-8723-e53ef109cc70-catalog-content\") pod \"redhat-operators-rfsn8\" (UID: \"2b33ed49-adf4-4398-8723-e53ef109cc70\") " pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.839324 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b33ed49-adf4-4398-8723-e53ef109cc70-utilities\") pod \"redhat-operators-rfsn8\" (UID: \"2b33ed49-adf4-4398-8723-e53ef109cc70\") " pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.839383 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c4wq\" (UniqueName: \"kubernetes.io/projected/2b33ed49-adf4-4398-8723-e53ef109cc70-kube-api-access-4c4wq\") pod \"redhat-operators-rfsn8\" (UID: \"2b33ed49-adf4-4398-8723-e53ef109cc70\") " pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.840114 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b33ed49-adf4-4398-8723-e53ef109cc70-catalog-content\") pod \"redhat-operators-rfsn8\" (UID: \"2b33ed49-adf4-4398-8723-e53ef109cc70\") " pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.840327 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b33ed49-adf4-4398-8723-e53ef109cc70-utilities\") pod \"redhat-operators-rfsn8\" (UID: \"2b33ed49-adf4-4398-8723-e53ef109cc70\") " pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.863069 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c4wq\" (UniqueName: \"kubernetes.io/projected/2b33ed49-adf4-4398-8723-e53ef109cc70-kube-api-access-4c4wq\") pod \"redhat-operators-rfsn8\" (UID: \"2b33ed49-adf4-4398-8723-e53ef109cc70\") " pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:42 crc kubenswrapper[4944]: I1124 09:15:42.947490 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:43 crc kubenswrapper[4944]: I1124 09:15:43.177185 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rfsn8"] Nov 24 09:15:44 crc kubenswrapper[4944]: I1124 09:15:44.198338 4944 generic.go:334] "Generic (PLEG): container finished" podID="2b33ed49-adf4-4398-8723-e53ef109cc70" containerID="2fdef0f78949953cbfbb92468f912719de65f291b17eed5f30903cf1976be391" exitCode=0 Nov 24 09:15:44 crc kubenswrapper[4944]: I1124 09:15:44.198481 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rfsn8" event={"ID":"2b33ed49-adf4-4398-8723-e53ef109cc70","Type":"ContainerDied","Data":"2fdef0f78949953cbfbb92468f912719de65f291b17eed5f30903cf1976be391"} Nov 24 09:15:44 crc kubenswrapper[4944]: I1124 09:15:44.200498 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rfsn8" event={"ID":"2b33ed49-adf4-4398-8723-e53ef109cc70","Type":"ContainerStarted","Data":"c4ce8fba8a48e44d44eced0d75e69bdf54afa64d7b3038da917097bd62ba2f41"} Nov 24 09:15:44 crc kubenswrapper[4944]: I1124 09:15:44.200255 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:15:51 crc kubenswrapper[4944]: I1124 09:15:51.260591 4944 generic.go:334] "Generic (PLEG): container finished" podID="2b33ed49-adf4-4398-8723-e53ef109cc70" containerID="e88e9687096a8e5a9bc1327b4a2bc76abdd1d349b21f01661211c0cfa5efa525" exitCode=0 Nov 24 09:15:51 crc kubenswrapper[4944]: I1124 09:15:51.260658 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rfsn8" event={"ID":"2b33ed49-adf4-4398-8723-e53ef109cc70","Type":"ContainerDied","Data":"e88e9687096a8e5a9bc1327b4a2bc76abdd1d349b21f01661211c0cfa5efa525"} Nov 24 09:15:52 crc kubenswrapper[4944]: I1124 09:15:52.272692 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rfsn8" event={"ID":"2b33ed49-adf4-4398-8723-e53ef109cc70","Type":"ContainerStarted","Data":"20de75456d2a8d7d2b8854bddbc229cbbd99da5df00727c93c1358c0bee99dcc"} Nov 24 09:15:52 crc kubenswrapper[4944]: I1124 09:15:52.292772 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rfsn8" podStartSLOduration=2.777183525 podStartE2EDuration="10.292755277s" podCreationTimestamp="2025-11-24 09:15:42 +0000 UTC" firstStartedPulling="2025-11-24 09:15:44.200012581 +0000 UTC m=+1404.734453043" lastFinishedPulling="2025-11-24 09:15:51.715584333 +0000 UTC m=+1412.250024795" observedRunningTime="2025-11-24 09:15:52.286769986 +0000 UTC m=+1412.821210468" watchObservedRunningTime="2025-11-24 09:15:52.292755277 +0000 UTC m=+1412.827195739" Nov 24 09:15:52 crc kubenswrapper[4944]: I1124 09:15:52.948454 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:52 crc kubenswrapper[4944]: I1124 09:15:52.948512 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:15:53 crc kubenswrapper[4944]: I1124 09:15:53.990823 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rfsn8" podUID="2b33ed49-adf4-4398-8723-e53ef109cc70" containerName="registry-server" probeResult="failure" output=< Nov 24 09:15:53 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 09:15:53 crc kubenswrapper[4944]: > Nov 24 09:16:02 crc kubenswrapper[4944]: I1124 09:16:02.987670 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.029718 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rfsn8" Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.099395 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rfsn8"] Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.221381 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kq7m4"] Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.221629 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kq7m4" podUID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" containerName="registry-server" containerID="cri-o://c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0" gracePeriod=2 Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.672136 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.847625 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xw9x\" (UniqueName: \"kubernetes.io/projected/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-kube-api-access-9xw9x\") pod \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.847680 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-catalog-content\") pod \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.847764 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-utilities\") pod \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\" (UID: \"1a5adcc4-e667-43f1-b5d5-fdca62ceced7\") " Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.848384 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-utilities" (OuterVolumeSpecName: "utilities") pod "1a5adcc4-e667-43f1-b5d5-fdca62ceced7" (UID: "1a5adcc4-e667-43f1-b5d5-fdca62ceced7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.854440 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-kube-api-access-9xw9x" (OuterVolumeSpecName: "kube-api-access-9xw9x") pod "1a5adcc4-e667-43f1-b5d5-fdca62ceced7" (UID: "1a5adcc4-e667-43f1-b5d5-fdca62ceced7"). InnerVolumeSpecName "kube-api-access-9xw9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.924872 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a5adcc4-e667-43f1-b5d5-fdca62ceced7" (UID: "1a5adcc4-e667-43f1-b5d5-fdca62ceced7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.949355 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xw9x\" (UniqueName: \"kubernetes.io/projected/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-kube-api-access-9xw9x\") on node \"crc\" DevicePath \"\"" Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.949388 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:16:03 crc kubenswrapper[4944]: I1124 09:16:03.949397 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a5adcc4-e667-43f1-b5d5-fdca62ceced7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.370027 4944 generic.go:334] "Generic (PLEG): container finished" podID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" containerID="c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0" exitCode=0 Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.370120 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq7m4" Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.370175 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7m4" event={"ID":"1a5adcc4-e667-43f1-b5d5-fdca62ceced7","Type":"ContainerDied","Data":"c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0"} Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.370213 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7m4" event={"ID":"1a5adcc4-e667-43f1-b5d5-fdca62ceced7","Type":"ContainerDied","Data":"17da6a16389fa2e0896a204ec4894b9d28e23a85ed87fed277cf75b79a665651"} Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.370234 4944 scope.go:117] "RemoveContainer" containerID="c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0" Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.397469 4944 scope.go:117] "RemoveContainer" containerID="19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec" Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.399429 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kq7m4"] Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.406941 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kq7m4"] Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.426141 4944 scope.go:117] "RemoveContainer" containerID="8f10f230ae51b931afaa7a5291e7ea55ddc33bf29364cc9278d5eb82b4957e3e" Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.450616 4944 scope.go:117] "RemoveContainer" containerID="c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0" Nov 24 09:16:04 crc kubenswrapper[4944]: E1124 09:16:04.451179 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0\": container with ID starting with c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0 not found: ID does not exist" containerID="c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0" Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.451268 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0"} err="failed to get container status \"c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0\": rpc error: code = NotFound desc = could not find container \"c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0\": container with ID starting with c9a2566243ae44a298b2a1c072b065141bfddeae16aa098d6d5b95f4ada9dcc0 not found: ID does not exist" Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.451315 4944 scope.go:117] "RemoveContainer" containerID="19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec" Nov 24 09:16:04 crc kubenswrapper[4944]: E1124 09:16:04.451749 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec\": container with ID starting with 19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec not found: ID does not exist" containerID="19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec" Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.451814 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec"} err="failed to get container status \"19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec\": rpc error: code = NotFound desc = could not find container \"19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec\": container with ID starting with 19f5b022487a596b4de98b3bd5a6f1df7ffd16c3a6377b2b989b8aa99a2b24ec not found: ID does not exist" Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.451841 4944 scope.go:117] "RemoveContainer" containerID="8f10f230ae51b931afaa7a5291e7ea55ddc33bf29364cc9278d5eb82b4957e3e" Nov 24 09:16:04 crc kubenswrapper[4944]: E1124 09:16:04.452161 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f10f230ae51b931afaa7a5291e7ea55ddc33bf29364cc9278d5eb82b4957e3e\": container with ID starting with 8f10f230ae51b931afaa7a5291e7ea55ddc33bf29364cc9278d5eb82b4957e3e not found: ID does not exist" containerID="8f10f230ae51b931afaa7a5291e7ea55ddc33bf29364cc9278d5eb82b4957e3e" Nov 24 09:16:04 crc kubenswrapper[4944]: I1124 09:16:04.452210 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f10f230ae51b931afaa7a5291e7ea55ddc33bf29364cc9278d5eb82b4957e3e"} err="failed to get container status \"8f10f230ae51b931afaa7a5291e7ea55ddc33bf29364cc9278d5eb82b4957e3e\": rpc error: code = NotFound desc = could not find container \"8f10f230ae51b931afaa7a5291e7ea55ddc33bf29364cc9278d5eb82b4957e3e\": container with ID starting with 8f10f230ae51b931afaa7a5291e7ea55ddc33bf29364cc9278d5eb82b4957e3e not found: ID does not exist" Nov 24 09:16:06 crc kubenswrapper[4944]: I1124 09:16:06.285133 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" path="/var/lib/kubelet/pods/1a5adcc4-e667-43f1-b5d5-fdca62ceced7/volumes" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.032845 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-drzp5"] Nov 24 09:16:07 crc kubenswrapper[4944]: E1124 09:16:07.033593 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" containerName="registry-server" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.033668 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" containerName="registry-server" Nov 24 09:16:07 crc kubenswrapper[4944]: E1124 09:16:07.033727 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" containerName="extract-utilities" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.033779 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" containerName="extract-utilities" Nov 24 09:16:07 crc kubenswrapper[4944]: E1124 09:16:07.033845 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" containerName="extract-content" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.033904 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" containerName="extract-content" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.034156 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a5adcc4-e667-43f1-b5d5-fdca62ceced7" containerName="registry-server" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.035300 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.047017 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-drzp5"] Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.194633 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-utilities\") pod \"certified-operators-drzp5\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.194951 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q75w9\" (UniqueName: \"kubernetes.io/projected/5acda24c-f409-4da4-8c51-754e4dbe23d7-kube-api-access-q75w9\") pod \"certified-operators-drzp5\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.195170 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-catalog-content\") pod \"certified-operators-drzp5\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.296489 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q75w9\" (UniqueName: \"kubernetes.io/projected/5acda24c-f409-4da4-8c51-754e4dbe23d7-kube-api-access-q75w9\") pod \"certified-operators-drzp5\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.296563 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-catalog-content\") pod \"certified-operators-drzp5\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.296621 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-utilities\") pod \"certified-operators-drzp5\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.297217 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-utilities\") pod \"certified-operators-drzp5\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.297344 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-catalog-content\") pod \"certified-operators-drzp5\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.323199 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q75w9\" (UniqueName: \"kubernetes.io/projected/5acda24c-f409-4da4-8c51-754e4dbe23d7-kube-api-access-q75w9\") pod \"certified-operators-drzp5\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.356644 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:07 crc kubenswrapper[4944]: I1124 09:16:07.683132 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-drzp5"] Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.032994 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hndh5"] Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.035229 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.058672 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hndh5"] Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.210712 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-utilities\") pod \"redhat-marketplace-hndh5\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.210781 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-catalog-content\") pod \"redhat-marketplace-hndh5\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.210827 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr7fz\" (UniqueName: \"kubernetes.io/projected/49f56202-9e8e-438a-9321-a01e0cf57711-kube-api-access-hr7fz\") pod \"redhat-marketplace-hndh5\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.312027 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-utilities\") pod \"redhat-marketplace-hndh5\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.312127 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-catalog-content\") pod \"redhat-marketplace-hndh5\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.312168 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr7fz\" (UniqueName: \"kubernetes.io/projected/49f56202-9e8e-438a-9321-a01e0cf57711-kube-api-access-hr7fz\") pod \"redhat-marketplace-hndh5\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.312764 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-catalog-content\") pod \"redhat-marketplace-hndh5\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.313033 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-utilities\") pod \"redhat-marketplace-hndh5\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.338118 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr7fz\" (UniqueName: \"kubernetes.io/projected/49f56202-9e8e-438a-9321-a01e0cf57711-kube-api-access-hr7fz\") pod \"redhat-marketplace-hndh5\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.404223 4944 generic.go:334] "Generic (PLEG): container finished" podID="5acda24c-f409-4da4-8c51-754e4dbe23d7" containerID="396a6662c272136c565ca8b53eae0577e31f16dbc0d5d58c6a61727bfbc89e1c" exitCode=0 Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.404311 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drzp5" event={"ID":"5acda24c-f409-4da4-8c51-754e4dbe23d7","Type":"ContainerDied","Data":"396a6662c272136c565ca8b53eae0577e31f16dbc0d5d58c6a61727bfbc89e1c"} Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.404534 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drzp5" event={"ID":"5acda24c-f409-4da4-8c51-754e4dbe23d7","Type":"ContainerStarted","Data":"0ab892964618add57fa1761ce46be46230680e647f5ddf161e15128378ffccb6"} Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.412881 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:08 crc kubenswrapper[4944]: I1124 09:16:08.849712 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hndh5"] Nov 24 09:16:09 crc kubenswrapper[4944]: I1124 09:16:09.413916 4944 generic.go:334] "Generic (PLEG): container finished" podID="49f56202-9e8e-438a-9321-a01e0cf57711" containerID="0197817f57011df76cf5f1ff6a25b500e238477302ceb6d627603dec255602a9" exitCode=0 Nov 24 09:16:09 crc kubenswrapper[4944]: I1124 09:16:09.413969 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hndh5" event={"ID":"49f56202-9e8e-438a-9321-a01e0cf57711","Type":"ContainerDied","Data":"0197817f57011df76cf5f1ff6a25b500e238477302ceb6d627603dec255602a9"} Nov 24 09:16:09 crc kubenswrapper[4944]: I1124 09:16:09.414015 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hndh5" event={"ID":"49f56202-9e8e-438a-9321-a01e0cf57711","Type":"ContainerStarted","Data":"bfcad80a3fd06bf32fbfd5a2fe33a3382d0bd975b4b8531d67cb87bf05217cc1"} Nov 24 09:16:10 crc kubenswrapper[4944]: I1124 09:16:10.422252 4944 generic.go:334] "Generic (PLEG): container finished" podID="49f56202-9e8e-438a-9321-a01e0cf57711" containerID="279c4fe9974ddbd0f3a791a3b295007a1b11cb7f5151d531d70ac543f9257db8" exitCode=0 Nov 24 09:16:10 crc kubenswrapper[4944]: I1124 09:16:10.422345 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hndh5" event={"ID":"49f56202-9e8e-438a-9321-a01e0cf57711","Type":"ContainerDied","Data":"279c4fe9974ddbd0f3a791a3b295007a1b11cb7f5151d531d70ac543f9257db8"} Nov 24 09:16:10 crc kubenswrapper[4944]: I1124 09:16:10.425447 4944 generic.go:334] "Generic (PLEG): container finished" podID="5acda24c-f409-4da4-8c51-754e4dbe23d7" containerID="56492fdf79d262751600c65c2212cd7e98cd27e2ff9a0b30d9ccd0ed4a89349b" exitCode=0 Nov 24 09:16:10 crc kubenswrapper[4944]: I1124 09:16:10.425496 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drzp5" event={"ID":"5acda24c-f409-4da4-8c51-754e4dbe23d7","Type":"ContainerDied","Data":"56492fdf79d262751600c65c2212cd7e98cd27e2ff9a0b30d9ccd0ed4a89349b"} Nov 24 09:16:11 crc kubenswrapper[4944]: I1124 09:16:11.434895 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drzp5" event={"ID":"5acda24c-f409-4da4-8c51-754e4dbe23d7","Type":"ContainerStarted","Data":"e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f"} Nov 24 09:16:11 crc kubenswrapper[4944]: I1124 09:16:11.437461 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hndh5" event={"ID":"49f56202-9e8e-438a-9321-a01e0cf57711","Type":"ContainerStarted","Data":"bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445"} Nov 24 09:16:11 crc kubenswrapper[4944]: I1124 09:16:11.465007 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-drzp5" podStartSLOduration=2.037600247 podStartE2EDuration="4.464986078s" podCreationTimestamp="2025-11-24 09:16:07 +0000 UTC" firstStartedPulling="2025-11-24 09:16:08.405827368 +0000 UTC m=+1428.940267830" lastFinishedPulling="2025-11-24 09:16:10.833213199 +0000 UTC m=+1431.367653661" observedRunningTime="2025-11-24 09:16:11.456767075 +0000 UTC m=+1431.991207547" watchObservedRunningTime="2025-11-24 09:16:11.464986078 +0000 UTC m=+1431.999426540" Nov 24 09:16:11 crc kubenswrapper[4944]: I1124 09:16:11.486128 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hndh5" podStartSLOduration=2.108585961 podStartE2EDuration="3.486110892s" podCreationTimestamp="2025-11-24 09:16:08 +0000 UTC" firstStartedPulling="2025-11-24 09:16:09.415802673 +0000 UTC m=+1429.950243135" lastFinishedPulling="2025-11-24 09:16:10.793327604 +0000 UTC m=+1431.327768066" observedRunningTime="2025-11-24 09:16:11.482063654 +0000 UTC m=+1432.016504136" watchObservedRunningTime="2025-11-24 09:16:11.486110892 +0000 UTC m=+1432.020551344" Nov 24 09:16:17 crc kubenswrapper[4944]: I1124 09:16:17.357693 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:17 crc kubenswrapper[4944]: I1124 09:16:17.357757 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:17 crc kubenswrapper[4944]: I1124 09:16:17.402612 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:17 crc kubenswrapper[4944]: I1124 09:16:17.520149 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:17 crc kubenswrapper[4944]: I1124 09:16:17.821765 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-drzp5"] Nov 24 09:16:18 crc kubenswrapper[4944]: I1124 09:16:18.413953 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:18 crc kubenswrapper[4944]: I1124 09:16:18.414956 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:18 crc kubenswrapper[4944]: I1124 09:16:18.456300 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:18 crc kubenswrapper[4944]: I1124 09:16:18.544749 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:19 crc kubenswrapper[4944]: I1124 09:16:19.511440 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-drzp5" podUID="5acda24c-f409-4da4-8c51-754e4dbe23d7" containerName="registry-server" containerID="cri-o://e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f" gracePeriod=2 Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.223791 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hndh5"] Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.339568 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.492759 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-utilities\") pod \"5acda24c-f409-4da4-8c51-754e4dbe23d7\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.492858 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q75w9\" (UniqueName: \"kubernetes.io/projected/5acda24c-f409-4da4-8c51-754e4dbe23d7-kube-api-access-q75w9\") pod \"5acda24c-f409-4da4-8c51-754e4dbe23d7\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.492903 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-catalog-content\") pod \"5acda24c-f409-4da4-8c51-754e4dbe23d7\" (UID: \"5acda24c-f409-4da4-8c51-754e4dbe23d7\") " Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.493741 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-utilities" (OuterVolumeSpecName: "utilities") pod "5acda24c-f409-4da4-8c51-754e4dbe23d7" (UID: "5acda24c-f409-4da4-8c51-754e4dbe23d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.498229 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5acda24c-f409-4da4-8c51-754e4dbe23d7-kube-api-access-q75w9" (OuterVolumeSpecName: "kube-api-access-q75w9") pod "5acda24c-f409-4da4-8c51-754e4dbe23d7" (UID: "5acda24c-f409-4da4-8c51-754e4dbe23d7"). InnerVolumeSpecName "kube-api-access-q75w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.524275 4944 generic.go:334] "Generic (PLEG): container finished" podID="5acda24c-f409-4da4-8c51-754e4dbe23d7" containerID="e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f" exitCode=0 Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.525019 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-drzp5" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.525510 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drzp5" event={"ID":"5acda24c-f409-4da4-8c51-754e4dbe23d7","Type":"ContainerDied","Data":"e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f"} Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.525549 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drzp5" event={"ID":"5acda24c-f409-4da4-8c51-754e4dbe23d7","Type":"ContainerDied","Data":"0ab892964618add57fa1761ce46be46230680e647f5ddf161e15128378ffccb6"} Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.525571 4944 scope.go:117] "RemoveContainer" containerID="e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.544957 4944 scope.go:117] "RemoveContainer" containerID="56492fdf79d262751600c65c2212cd7e98cd27e2ff9a0b30d9ccd0ed4a89349b" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.548871 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5acda24c-f409-4da4-8c51-754e4dbe23d7" (UID: "5acda24c-f409-4da4-8c51-754e4dbe23d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.564658 4944 scope.go:117] "RemoveContainer" containerID="396a6662c272136c565ca8b53eae0577e31f16dbc0d5d58c6a61727bfbc89e1c" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.591674 4944 scope.go:117] "RemoveContainer" containerID="e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f" Nov 24 09:16:20 crc kubenswrapper[4944]: E1124 09:16:20.593525 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f\": container with ID starting with e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f not found: ID does not exist" containerID="e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.597550 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f"} err="failed to get container status \"e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f\": rpc error: code = NotFound desc = could not find container \"e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f\": container with ID starting with e74073fa5375c1bb1f0d0d09d3202330e0e6596fb4b385a727a8343dd967904f not found: ID does not exist" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.597839 4944 scope.go:117] "RemoveContainer" containerID="56492fdf79d262751600c65c2212cd7e98cd27e2ff9a0b30d9ccd0ed4a89349b" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.594529 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q75w9\" (UniqueName: \"kubernetes.io/projected/5acda24c-f409-4da4-8c51-754e4dbe23d7-kube-api-access-q75w9\") on node \"crc\" DevicePath \"\"" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.598069 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.598088 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5acda24c-f409-4da4-8c51-754e4dbe23d7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:16:20 crc kubenswrapper[4944]: E1124 09:16:20.599062 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56492fdf79d262751600c65c2212cd7e98cd27e2ff9a0b30d9ccd0ed4a89349b\": container with ID starting with 56492fdf79d262751600c65c2212cd7e98cd27e2ff9a0b30d9ccd0ed4a89349b not found: ID does not exist" containerID="56492fdf79d262751600c65c2212cd7e98cd27e2ff9a0b30d9ccd0ed4a89349b" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.599112 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56492fdf79d262751600c65c2212cd7e98cd27e2ff9a0b30d9ccd0ed4a89349b"} err="failed to get container status \"56492fdf79d262751600c65c2212cd7e98cd27e2ff9a0b30d9ccd0ed4a89349b\": rpc error: code = NotFound desc = could not find container \"56492fdf79d262751600c65c2212cd7e98cd27e2ff9a0b30d9ccd0ed4a89349b\": container with ID starting with 56492fdf79d262751600c65c2212cd7e98cd27e2ff9a0b30d9ccd0ed4a89349b not found: ID does not exist" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.599138 4944 scope.go:117] "RemoveContainer" containerID="396a6662c272136c565ca8b53eae0577e31f16dbc0d5d58c6a61727bfbc89e1c" Nov 24 09:16:20 crc kubenswrapper[4944]: E1124 09:16:20.600087 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"396a6662c272136c565ca8b53eae0577e31f16dbc0d5d58c6a61727bfbc89e1c\": container with ID starting with 396a6662c272136c565ca8b53eae0577e31f16dbc0d5d58c6a61727bfbc89e1c not found: ID does not exist" containerID="396a6662c272136c565ca8b53eae0577e31f16dbc0d5d58c6a61727bfbc89e1c" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.600115 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"396a6662c272136c565ca8b53eae0577e31f16dbc0d5d58c6a61727bfbc89e1c"} err="failed to get container status \"396a6662c272136c565ca8b53eae0577e31f16dbc0d5d58c6a61727bfbc89e1c\": rpc error: code = NotFound desc = could not find container \"396a6662c272136c565ca8b53eae0577e31f16dbc0d5d58c6a61727bfbc89e1c\": container with ID starting with 396a6662c272136c565ca8b53eae0577e31f16dbc0d5d58c6a61727bfbc89e1c not found: ID does not exist" Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.855772 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-drzp5"] Nov 24 09:16:20 crc kubenswrapper[4944]: I1124 09:16:20.861119 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-drzp5"] Nov 24 09:16:21 crc kubenswrapper[4944]: I1124 09:16:21.538689 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hndh5" podUID="49f56202-9e8e-438a-9321-a01e0cf57711" containerName="registry-server" containerID="cri-o://bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445" gracePeriod=2 Nov 24 09:16:21 crc kubenswrapper[4944]: I1124 09:16:21.910774 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.018570 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-catalog-content\") pod \"49f56202-9e8e-438a-9321-a01e0cf57711\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.018663 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr7fz\" (UniqueName: \"kubernetes.io/projected/49f56202-9e8e-438a-9321-a01e0cf57711-kube-api-access-hr7fz\") pod \"49f56202-9e8e-438a-9321-a01e0cf57711\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.018825 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-utilities\") pod \"49f56202-9e8e-438a-9321-a01e0cf57711\" (UID: \"49f56202-9e8e-438a-9321-a01e0cf57711\") " Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.019636 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-utilities" (OuterVolumeSpecName: "utilities") pod "49f56202-9e8e-438a-9321-a01e0cf57711" (UID: "49f56202-9e8e-438a-9321-a01e0cf57711"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.023545 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49f56202-9e8e-438a-9321-a01e0cf57711-kube-api-access-hr7fz" (OuterVolumeSpecName: "kube-api-access-hr7fz") pod "49f56202-9e8e-438a-9321-a01e0cf57711" (UID: "49f56202-9e8e-438a-9321-a01e0cf57711"). InnerVolumeSpecName "kube-api-access-hr7fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.037969 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49f56202-9e8e-438a-9321-a01e0cf57711" (UID: "49f56202-9e8e-438a-9321-a01e0cf57711"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.120406 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.120445 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr7fz\" (UniqueName: \"kubernetes.io/projected/49f56202-9e8e-438a-9321-a01e0cf57711-kube-api-access-hr7fz\") on node \"crc\" DevicePath \"\"" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.120458 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49f56202-9e8e-438a-9321-a01e0cf57711-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.284747 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5acda24c-f409-4da4-8c51-754e4dbe23d7" path="/var/lib/kubelet/pods/5acda24c-f409-4da4-8c51-754e4dbe23d7/volumes" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.560466 4944 generic.go:334] "Generic (PLEG): container finished" podID="49f56202-9e8e-438a-9321-a01e0cf57711" containerID="bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445" exitCode=0 Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.560524 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hndh5" event={"ID":"49f56202-9e8e-438a-9321-a01e0cf57711","Type":"ContainerDied","Data":"bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445"} Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.560560 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hndh5" event={"ID":"49f56202-9e8e-438a-9321-a01e0cf57711","Type":"ContainerDied","Data":"bfcad80a3fd06bf32fbfd5a2fe33a3382d0bd975b4b8531d67cb87bf05217cc1"} Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.560613 4944 scope.go:117] "RemoveContainer" containerID="bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.560893 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hndh5" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.587709 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hndh5"] Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.592273 4944 scope.go:117] "RemoveContainer" containerID="279c4fe9974ddbd0f3a791a3b295007a1b11cb7f5151d531d70ac543f9257db8" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.592601 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hndh5"] Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.615271 4944 scope.go:117] "RemoveContainer" containerID="0197817f57011df76cf5f1ff6a25b500e238477302ceb6d627603dec255602a9" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.643705 4944 scope.go:117] "RemoveContainer" containerID="bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445" Nov 24 09:16:22 crc kubenswrapper[4944]: E1124 09:16:22.644359 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445\": container with ID starting with bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445 not found: ID does not exist" containerID="bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.644417 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445"} err="failed to get container status \"bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445\": rpc error: code = NotFound desc = could not find container \"bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445\": container with ID starting with bdd5205c32546f8f982aee76d08de2313582c2ecec2d1330db14c8b3e7add445 not found: ID does not exist" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.644455 4944 scope.go:117] "RemoveContainer" containerID="279c4fe9974ddbd0f3a791a3b295007a1b11cb7f5151d531d70ac543f9257db8" Nov 24 09:16:22 crc kubenswrapper[4944]: E1124 09:16:22.644841 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"279c4fe9974ddbd0f3a791a3b295007a1b11cb7f5151d531d70ac543f9257db8\": container with ID starting with 279c4fe9974ddbd0f3a791a3b295007a1b11cb7f5151d531d70ac543f9257db8 not found: ID does not exist" containerID="279c4fe9974ddbd0f3a791a3b295007a1b11cb7f5151d531d70ac543f9257db8" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.644871 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"279c4fe9974ddbd0f3a791a3b295007a1b11cb7f5151d531d70ac543f9257db8"} err="failed to get container status \"279c4fe9974ddbd0f3a791a3b295007a1b11cb7f5151d531d70ac543f9257db8\": rpc error: code = NotFound desc = could not find container \"279c4fe9974ddbd0f3a791a3b295007a1b11cb7f5151d531d70ac543f9257db8\": container with ID starting with 279c4fe9974ddbd0f3a791a3b295007a1b11cb7f5151d531d70ac543f9257db8 not found: ID does not exist" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.644892 4944 scope.go:117] "RemoveContainer" containerID="0197817f57011df76cf5f1ff6a25b500e238477302ceb6d627603dec255602a9" Nov 24 09:16:22 crc kubenswrapper[4944]: E1124 09:16:22.645246 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0197817f57011df76cf5f1ff6a25b500e238477302ceb6d627603dec255602a9\": container with ID starting with 0197817f57011df76cf5f1ff6a25b500e238477302ceb6d627603dec255602a9 not found: ID does not exist" containerID="0197817f57011df76cf5f1ff6a25b500e238477302ceb6d627603dec255602a9" Nov 24 09:16:22 crc kubenswrapper[4944]: I1124 09:16:22.645287 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0197817f57011df76cf5f1ff6a25b500e238477302ceb6d627603dec255602a9"} err="failed to get container status \"0197817f57011df76cf5f1ff6a25b500e238477302ceb6d627603dec255602a9\": rpc error: code = NotFound desc = could not find container \"0197817f57011df76cf5f1ff6a25b500e238477302ceb6d627603dec255602a9\": container with ID starting with 0197817f57011df76cf5f1ff6a25b500e238477302ceb6d627603dec255602a9 not found: ID does not exist" Nov 24 09:16:24 crc kubenswrapper[4944]: I1124 09:16:24.287862 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49f56202-9e8e-438a-9321-a01e0cf57711" path="/var/lib/kubelet/pods/49f56202-9e8e-438a-9321-a01e0cf57711/volumes" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.287983 4944 scope.go:117] "RemoveContainer" containerID="c7910fb866d5b046b5b553762964b079616ae2b4bfdc94ec40aa3f1676ccbbac" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.313542 4944 scope.go:117] "RemoveContainer" containerID="35d72de3e7b54bd62e574fcc3531f7f21568189bde22bde766d7472b5f0e4bfb" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.339993 4944 scope.go:117] "RemoveContainer" containerID="6a3410c6257d72bcea62d39ce6013b2c346a2c2f1e11397a1e28d3f46511dc6e" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.363377 4944 scope.go:117] "RemoveContainer" containerID="99bf25824c378bb0feb5d33ed10f3a48b6d21f932f260473ccb0a6b7c666246d" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.389067 4944 scope.go:117] "RemoveContainer" containerID="86988b92220b8f00057d839eacd2728b2c903a679a7f8d2f627e8ac04be99bcd" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.427087 4944 scope.go:117] "RemoveContainer" containerID="de79cb55674ca0207fc7a59b27c802ecc027ac030257f5a33d496e3c2b02b09a" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.464623 4944 scope.go:117] "RemoveContainer" containerID="81af0f39619d4f0d4399fe088165e6c6fb343c03871f10888e1d787a30b8aa80" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.490180 4944 scope.go:117] "RemoveContainer" containerID="59887e3dd78e9b25b49b5075672aa68784184f607941ee0b247b3554533b3f0e" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.514077 4944 scope.go:117] "RemoveContainer" containerID="05f3827d7409a75ca27686ed4e2f7544dca88296220b6c81cd7eceee4395c8cb" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.534523 4944 scope.go:117] "RemoveContainer" containerID="3f8102c9e1131a4291d679c8ceadc0d3f525ccd386d483f44e28e313702eb963" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.551865 4944 scope.go:117] "RemoveContainer" containerID="66e04adc7715ce4a2e86d8053addf377b2b9540de52cbc2b0ed756013704a659" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.572390 4944 scope.go:117] "RemoveContainer" containerID="5e8769115bdfd3dfb5565693bdba51e398080bd70b2c634c3bca81f94b52896c" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.591683 4944 scope.go:117] "RemoveContainer" containerID="5f7f01bba3bbc83c216d9707bc817704d3d12a1909dfb8fada43a600b9229191" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.627126 4944 scope.go:117] "RemoveContainer" containerID="751b73b489bf6597417b5d9527e9febd188bf9c2d66821855286a3d83547ec5d" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.643902 4944 scope.go:117] "RemoveContainer" containerID="1bfe5877d95a636388b958f1f7d0d9a81c2b7c523ca095ed7bbdf4c983a6a8b6" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.664993 4944 scope.go:117] "RemoveContainer" containerID="0f0f741121c6d302e3545b02a6d4a19fa67ea57a811079fb4057d58abb0f5764" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.692534 4944 scope.go:117] "RemoveContainer" containerID="10473e2e58eff197e7d6054c78b1b43c310d54ae418a52fb6168c9c0b59133e8" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.711611 4944 scope.go:117] "RemoveContainer" containerID="e3ebe45315d452b608c18183ab8a4f20e94a8a7e1e85df1008d16d4287b10042" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.733321 4944 scope.go:117] "RemoveContainer" containerID="879f66b45dd1486ce94200c6d646a5c12a6b2404828999d8fa88dd82bb7f9229" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.751871 4944 scope.go:117] "RemoveContainer" containerID="88dc28611f87d15ce858b43d3b87af57347bf25a3db8b6f40bb1b9316cec873b" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.775869 4944 scope.go:117] "RemoveContainer" containerID="6d6ec38d97812dfd2d8aba09a7804fe182ba21ae83c5a722630f8851be56f865" Nov 24 09:16:26 crc kubenswrapper[4944]: I1124 09:16:26.800298 4944 scope.go:117] "RemoveContainer" containerID="0fe726f95d087770f3a55da5b8318319bcd580d793904dd02e38159fed22e981" Nov 24 09:16:53 crc kubenswrapper[4944]: I1124 09:16:53.548398 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:16:53 crc kubenswrapper[4944]: I1124 09:16:53.549039 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:17:23 crc kubenswrapper[4944]: I1124 09:17:23.548592 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:17:23 crc kubenswrapper[4944]: I1124 09:17:23.549114 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.137571 4944 scope.go:117] "RemoveContainer" containerID="026e0e21f1a64afd79436eb0f96f0119e4d149466ea95e71c69fb8bd412bf0e1" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.186991 4944 scope.go:117] "RemoveContainer" containerID="ceafa9e1a0d2708df2e25915e16d1a81cfc2d983d0d29259d0153aeb04d1d53f" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.212620 4944 scope.go:117] "RemoveContainer" containerID="0d336c6d00011b405b89e69c75750bd4892dffa8ff7440cdf41f4517d10dd129" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.241796 4944 scope.go:117] "RemoveContainer" containerID="60c6a295891b9eabbf6aff3ae56e1707132a7e25f4e5417b47601279a7b69df1" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.283891 4944 scope.go:117] "RemoveContainer" containerID="645b1d9673243c219c166fe53c247425c184e6160cec86a567ca59eeea550661" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.324547 4944 scope.go:117] "RemoveContainer" containerID="6f04f1e4430849952d1609d61a9811d5bc67da263bac5157d0480d29fa8bd109" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.362877 4944 scope.go:117] "RemoveContainer" containerID="e734c391470a420ba500ddc3044001dc8a942a79f4c8290fe6f87c79d611a796" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.389481 4944 scope.go:117] "RemoveContainer" containerID="5950163d1a4608641fb113f50f6737d5436fc15138c0f58c9d9ec29a06eda578" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.408471 4944 scope.go:117] "RemoveContainer" containerID="0c0565dec9e48700e3377ada79e525e8bfcaf6d1918e31822b21484621342dca" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.444003 4944 scope.go:117] "RemoveContainer" containerID="446e72ed669c7790b5e97ed81cfd34742c7c9da15ec2b51504d0e8b3f86995c1" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.477630 4944 scope.go:117] "RemoveContainer" containerID="32fb6228e84968d5e793ee6c46d782e11a2ef4eebf6c818f2226f27364162b50" Nov 24 09:17:27 crc kubenswrapper[4944]: I1124 09:17:27.495685 4944 scope.go:117] "RemoveContainer" containerID="4c6a05fbd86035dc80b848ecd34654a1e05f03fb73dc590f509e5bdbc7cb4bc2" Nov 24 09:17:53 crc kubenswrapper[4944]: I1124 09:17:53.548440 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:17:53 crc kubenswrapper[4944]: I1124 09:17:53.549035 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:17:53 crc kubenswrapper[4944]: I1124 09:17:53.549190 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:17:53 crc kubenswrapper[4944]: I1124 09:17:53.549741 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:17:53 crc kubenswrapper[4944]: I1124 09:17:53.549801 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" gracePeriod=600 Nov 24 09:17:53 crc kubenswrapper[4944]: E1124 09:17:53.667918 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:17:54 crc kubenswrapper[4944]: I1124 09:17:54.240174 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" exitCode=0 Nov 24 09:17:54 crc kubenswrapper[4944]: I1124 09:17:54.240215 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731"} Nov 24 09:17:54 crc kubenswrapper[4944]: I1124 09:17:54.240248 4944 scope.go:117] "RemoveContainer" containerID="f8ac17962023518b7abc2c1aa6e647faf7d9a5693323d499415d0670598b6a65" Nov 24 09:17:54 crc kubenswrapper[4944]: I1124 09:17:54.241021 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:17:54 crc kubenswrapper[4944]: E1124 09:17:54.241439 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:18:05 crc kubenswrapper[4944]: I1124 09:18:05.276984 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:18:05 crc kubenswrapper[4944]: E1124 09:18:05.277759 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:18:19 crc kubenswrapper[4944]: I1124 09:18:19.276816 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:18:19 crc kubenswrapper[4944]: E1124 09:18:19.278575 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:18:27 crc kubenswrapper[4944]: I1124 09:18:27.648295 4944 scope.go:117] "RemoveContainer" containerID="8ee523b44e35b1cfe2ffff334cb3b42352d07e0931e975529ab8ff087be0b252" Nov 24 09:18:27 crc kubenswrapper[4944]: I1124 09:18:27.681084 4944 scope.go:117] "RemoveContainer" containerID="5c74c827e2500bbe5cec629901b4499f9820b0c3015d1b7fc182c6d47192e91d" Nov 24 09:18:27 crc kubenswrapper[4944]: I1124 09:18:27.706518 4944 scope.go:117] "RemoveContainer" containerID="8e5045f0585e8dd46160e5e97aae1809dc8d19296f6529d8d76bb17c76758990" Nov 24 09:18:27 crc kubenswrapper[4944]: I1124 09:18:27.740092 4944 scope.go:117] "RemoveContainer" containerID="194effb1be878a12474f5aad321c9b034378e690429268d01ba3188c16709667" Nov 24 09:18:27 crc kubenswrapper[4944]: I1124 09:18:27.766104 4944 scope.go:117] "RemoveContainer" containerID="da767b6a401b73c320d8ff35a4d71eba18485e8bfac8bc5be97a2e7215f5f07b" Nov 24 09:18:27 crc kubenswrapper[4944]: I1124 09:18:27.795756 4944 scope.go:117] "RemoveContainer" containerID="b626ffc5d738486bd96a728516fb7e89c1bf6800fb2f82fd0b1bc9e2f6d19268" Nov 24 09:18:27 crc kubenswrapper[4944]: I1124 09:18:27.814692 4944 scope.go:117] "RemoveContainer" containerID="7fb724f45219e9d1f3b1a09039899d4d0f3f225aabff2cb17b1ae874531f5a97" Nov 24 09:18:27 crc kubenswrapper[4944]: I1124 09:18:27.831355 4944 scope.go:117] "RemoveContainer" containerID="2c1a579178840a65cb60ad40f3b83a1c327413e6a2d77e8178e33176b65db789" Nov 24 09:18:27 crc kubenswrapper[4944]: I1124 09:18:27.866250 4944 scope.go:117] "RemoveContainer" containerID="38bd22ea8e5f5357cd4719266b4150ee62593a72993379b83a9ba97d0c889479" Nov 24 09:18:27 crc kubenswrapper[4944]: I1124 09:18:27.890063 4944 scope.go:117] "RemoveContainer" containerID="ba5de3bc08fb3adcb30a032e46fa2306b0ea56b6eef98ca20a635459d7ab466d" Nov 24 09:18:33 crc kubenswrapper[4944]: I1124 09:18:33.277509 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:18:33 crc kubenswrapper[4944]: E1124 09:18:33.278894 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:18:45 crc kubenswrapper[4944]: I1124 09:18:45.276861 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:18:45 crc kubenswrapper[4944]: E1124 09:18:45.277655 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:18:57 crc kubenswrapper[4944]: I1124 09:18:57.276626 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:18:57 crc kubenswrapper[4944]: E1124 09:18:57.277386 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:19:08 crc kubenswrapper[4944]: I1124 09:19:08.276905 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:19:08 crc kubenswrapper[4944]: E1124 09:19:08.277630 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:19:21 crc kubenswrapper[4944]: I1124 09:19:21.276243 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:19:21 crc kubenswrapper[4944]: E1124 09:19:21.276962 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:19:28 crc kubenswrapper[4944]: I1124 09:19:28.007455 4944 scope.go:117] "RemoveContainer" containerID="4e421ffe1b1f553a1d508dca1e10832265fc9517f18241ac1ae0a844024768cc" Nov 24 09:19:28 crc kubenswrapper[4944]: I1124 09:19:28.054181 4944 scope.go:117] "RemoveContainer" containerID="afb36cdea5b1e686a6adc178f7749637e22fb3c8332fe0b88d9e35de15895874" Nov 24 09:19:34 crc kubenswrapper[4944]: I1124 09:19:34.277041 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:19:34 crc kubenswrapper[4944]: E1124 09:19:34.277770 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:19:49 crc kubenswrapper[4944]: I1124 09:19:49.276719 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:19:49 crc kubenswrapper[4944]: E1124 09:19:49.277469 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:20:00 crc kubenswrapper[4944]: I1124 09:20:00.281678 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:20:00 crc kubenswrapper[4944]: E1124 09:20:00.283304 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.353242 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tnkjz"] Nov 24 09:20:08 crc kubenswrapper[4944]: E1124 09:20:08.355633 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49f56202-9e8e-438a-9321-a01e0cf57711" containerName="registry-server" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.355654 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="49f56202-9e8e-438a-9321-a01e0cf57711" containerName="registry-server" Nov 24 09:20:08 crc kubenswrapper[4944]: E1124 09:20:08.355674 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5acda24c-f409-4da4-8c51-754e4dbe23d7" containerName="extract-utilities" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.355682 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5acda24c-f409-4da4-8c51-754e4dbe23d7" containerName="extract-utilities" Nov 24 09:20:08 crc kubenswrapper[4944]: E1124 09:20:08.355693 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49f56202-9e8e-438a-9321-a01e0cf57711" containerName="extract-utilities" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.355699 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="49f56202-9e8e-438a-9321-a01e0cf57711" containerName="extract-utilities" Nov 24 09:20:08 crc kubenswrapper[4944]: E1124 09:20:08.355711 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5acda24c-f409-4da4-8c51-754e4dbe23d7" containerName="registry-server" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.355717 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5acda24c-f409-4da4-8c51-754e4dbe23d7" containerName="registry-server" Nov 24 09:20:08 crc kubenswrapper[4944]: E1124 09:20:08.355724 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5acda24c-f409-4da4-8c51-754e4dbe23d7" containerName="extract-content" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.355731 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5acda24c-f409-4da4-8c51-754e4dbe23d7" containerName="extract-content" Nov 24 09:20:08 crc kubenswrapper[4944]: E1124 09:20:08.355744 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49f56202-9e8e-438a-9321-a01e0cf57711" containerName="extract-content" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.355750 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="49f56202-9e8e-438a-9321-a01e0cf57711" containerName="extract-content" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.355917 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="49f56202-9e8e-438a-9321-a01e0cf57711" containerName="registry-server" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.355937 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5acda24c-f409-4da4-8c51-754e4dbe23d7" containerName="registry-server" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.356900 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.367393 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tnkjz"] Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.459125 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-utilities\") pod \"community-operators-tnkjz\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.459507 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-catalog-content\") pod \"community-operators-tnkjz\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.459549 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sklm9\" (UniqueName: \"kubernetes.io/projected/0e28f479-229b-4d11-8d45-32b16edcc808-kube-api-access-sklm9\") pod \"community-operators-tnkjz\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.560465 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-catalog-content\") pod \"community-operators-tnkjz\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.560548 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sklm9\" (UniqueName: \"kubernetes.io/projected/0e28f479-229b-4d11-8d45-32b16edcc808-kube-api-access-sklm9\") pod \"community-operators-tnkjz\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.560921 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-catalog-content\") pod \"community-operators-tnkjz\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.561295 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-utilities\") pod \"community-operators-tnkjz\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.561008 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-utilities\") pod \"community-operators-tnkjz\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.579347 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sklm9\" (UniqueName: \"kubernetes.io/projected/0e28f479-229b-4d11-8d45-32b16edcc808-kube-api-access-sklm9\") pod \"community-operators-tnkjz\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.686421 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:08 crc kubenswrapper[4944]: I1124 09:20:08.942021 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tnkjz"] Nov 24 09:20:09 crc kubenswrapper[4944]: I1124 09:20:09.205253 4944 generic.go:334] "Generic (PLEG): container finished" podID="0e28f479-229b-4d11-8d45-32b16edcc808" containerID="15fe300a40c28e63c0f2d9e717b4035c2926f52744ad29a70b0f5d84bae0a60d" exitCode=0 Nov 24 09:20:09 crc kubenswrapper[4944]: I1124 09:20:09.205360 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnkjz" event={"ID":"0e28f479-229b-4d11-8d45-32b16edcc808","Type":"ContainerDied","Data":"15fe300a40c28e63c0f2d9e717b4035c2926f52744ad29a70b0f5d84bae0a60d"} Nov 24 09:20:09 crc kubenswrapper[4944]: I1124 09:20:09.205932 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnkjz" event={"ID":"0e28f479-229b-4d11-8d45-32b16edcc808","Type":"ContainerStarted","Data":"84240a2e7326053cd3d5818742d0c25d686bcc93217636ebcc60a2cd362f4529"} Nov 24 09:20:10 crc kubenswrapper[4944]: I1124 09:20:10.219568 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnkjz" event={"ID":"0e28f479-229b-4d11-8d45-32b16edcc808","Type":"ContainerStarted","Data":"69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3"} Nov 24 09:20:11 crc kubenswrapper[4944]: I1124 09:20:11.227613 4944 generic.go:334] "Generic (PLEG): container finished" podID="0e28f479-229b-4d11-8d45-32b16edcc808" containerID="69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3" exitCode=0 Nov 24 09:20:11 crc kubenswrapper[4944]: I1124 09:20:11.227656 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnkjz" event={"ID":"0e28f479-229b-4d11-8d45-32b16edcc808","Type":"ContainerDied","Data":"69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3"} Nov 24 09:20:12 crc kubenswrapper[4944]: I1124 09:20:12.248070 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnkjz" event={"ID":"0e28f479-229b-4d11-8d45-32b16edcc808","Type":"ContainerStarted","Data":"febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728"} Nov 24 09:20:12 crc kubenswrapper[4944]: I1124 09:20:12.270103 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tnkjz" podStartSLOduration=1.7011577070000001 podStartE2EDuration="4.270081834s" podCreationTimestamp="2025-11-24 09:20:08 +0000 UTC" firstStartedPulling="2025-11-24 09:20:09.210056657 +0000 UTC m=+1669.744497129" lastFinishedPulling="2025-11-24 09:20:11.778980794 +0000 UTC m=+1672.313421256" observedRunningTime="2025-11-24 09:20:12.265143017 +0000 UTC m=+1672.799583479" watchObservedRunningTime="2025-11-24 09:20:12.270081834 +0000 UTC m=+1672.804522296" Nov 24 09:20:14 crc kubenswrapper[4944]: I1124 09:20:14.277750 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:20:14 crc kubenswrapper[4944]: E1124 09:20:14.278097 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:20:18 crc kubenswrapper[4944]: I1124 09:20:18.687137 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:18 crc kubenswrapper[4944]: I1124 09:20:18.688225 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:18 crc kubenswrapper[4944]: I1124 09:20:18.734621 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:19 crc kubenswrapper[4944]: I1124 09:20:19.352152 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:19 crc kubenswrapper[4944]: I1124 09:20:19.396843 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tnkjz"] Nov 24 09:20:21 crc kubenswrapper[4944]: I1124 09:20:21.311393 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tnkjz" podUID="0e28f479-229b-4d11-8d45-32b16edcc808" containerName="registry-server" containerID="cri-o://febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728" gracePeriod=2 Nov 24 09:20:21 crc kubenswrapper[4944]: I1124 09:20:21.676196 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:21 crc kubenswrapper[4944]: I1124 09:20:21.846365 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-catalog-content\") pod \"0e28f479-229b-4d11-8d45-32b16edcc808\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " Nov 24 09:20:21 crc kubenswrapper[4944]: I1124 09:20:21.846824 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-utilities\") pod \"0e28f479-229b-4d11-8d45-32b16edcc808\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " Nov 24 09:20:21 crc kubenswrapper[4944]: I1124 09:20:21.846939 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sklm9\" (UniqueName: \"kubernetes.io/projected/0e28f479-229b-4d11-8d45-32b16edcc808-kube-api-access-sklm9\") pod \"0e28f479-229b-4d11-8d45-32b16edcc808\" (UID: \"0e28f479-229b-4d11-8d45-32b16edcc808\") " Nov 24 09:20:21 crc kubenswrapper[4944]: I1124 09:20:21.852910 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e28f479-229b-4d11-8d45-32b16edcc808-kube-api-access-sklm9" (OuterVolumeSpecName: "kube-api-access-sklm9") pod "0e28f479-229b-4d11-8d45-32b16edcc808" (UID: "0e28f479-229b-4d11-8d45-32b16edcc808"). InnerVolumeSpecName "kube-api-access-sklm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:20:21 crc kubenswrapper[4944]: I1124 09:20:21.857941 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-utilities" (OuterVolumeSpecName: "utilities") pod "0e28f479-229b-4d11-8d45-32b16edcc808" (UID: "0e28f479-229b-4d11-8d45-32b16edcc808"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:20:21 crc kubenswrapper[4944]: I1124 09:20:21.905539 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e28f479-229b-4d11-8d45-32b16edcc808" (UID: "0e28f479-229b-4d11-8d45-32b16edcc808"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:20:21 crc kubenswrapper[4944]: I1124 09:20:21.949086 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:21 crc kubenswrapper[4944]: I1124 09:20:21.949123 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sklm9\" (UniqueName: \"kubernetes.io/projected/0e28f479-229b-4d11-8d45-32b16edcc808-kube-api-access-sklm9\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:21 crc kubenswrapper[4944]: I1124 09:20:21.949134 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e28f479-229b-4d11-8d45-32b16edcc808-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.320542 4944 generic.go:334] "Generic (PLEG): container finished" podID="0e28f479-229b-4d11-8d45-32b16edcc808" containerID="febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728" exitCode=0 Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.320583 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnkjz" event={"ID":"0e28f479-229b-4d11-8d45-32b16edcc808","Type":"ContainerDied","Data":"febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728"} Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.320608 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnkjz" event={"ID":"0e28f479-229b-4d11-8d45-32b16edcc808","Type":"ContainerDied","Data":"84240a2e7326053cd3d5818742d0c25d686bcc93217636ebcc60a2cd362f4529"} Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.320628 4944 scope.go:117] "RemoveContainer" containerID="febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728" Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.320647 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnkjz" Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.346895 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tnkjz"] Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.353370 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tnkjz"] Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.356139 4944 scope.go:117] "RemoveContainer" containerID="69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3" Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.373686 4944 scope.go:117] "RemoveContainer" containerID="15fe300a40c28e63c0f2d9e717b4035c2926f52744ad29a70b0f5d84bae0a60d" Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.394155 4944 scope.go:117] "RemoveContainer" containerID="febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728" Nov 24 09:20:22 crc kubenswrapper[4944]: E1124 09:20:22.394653 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728\": container with ID starting with febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728 not found: ID does not exist" containerID="febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728" Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.394708 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728"} err="failed to get container status \"febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728\": rpc error: code = NotFound desc = could not find container \"febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728\": container with ID starting with febd109943509ec44d722a1e9723d347fe875785044bedddd0b0672f176cf728 not found: ID does not exist" Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.394732 4944 scope.go:117] "RemoveContainer" containerID="69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3" Nov 24 09:20:22 crc kubenswrapper[4944]: E1124 09:20:22.395017 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3\": container with ID starting with 69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3 not found: ID does not exist" containerID="69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3" Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.395100 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3"} err="failed to get container status \"69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3\": rpc error: code = NotFound desc = could not find container \"69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3\": container with ID starting with 69ceb1790564989fcfa29603ca3d169a37e04da8442f4e59c14d2b938c3f62c3 not found: ID does not exist" Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.395133 4944 scope.go:117] "RemoveContainer" containerID="15fe300a40c28e63c0f2d9e717b4035c2926f52744ad29a70b0f5d84bae0a60d" Nov 24 09:20:22 crc kubenswrapper[4944]: E1124 09:20:22.395532 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15fe300a40c28e63c0f2d9e717b4035c2926f52744ad29a70b0f5d84bae0a60d\": container with ID starting with 15fe300a40c28e63c0f2d9e717b4035c2926f52744ad29a70b0f5d84bae0a60d not found: ID does not exist" containerID="15fe300a40c28e63c0f2d9e717b4035c2926f52744ad29a70b0f5d84bae0a60d" Nov 24 09:20:22 crc kubenswrapper[4944]: I1124 09:20:22.395558 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15fe300a40c28e63c0f2d9e717b4035c2926f52744ad29a70b0f5d84bae0a60d"} err="failed to get container status \"15fe300a40c28e63c0f2d9e717b4035c2926f52744ad29a70b0f5d84bae0a60d\": rpc error: code = NotFound desc = could not find container \"15fe300a40c28e63c0f2d9e717b4035c2926f52744ad29a70b0f5d84bae0a60d\": container with ID starting with 15fe300a40c28e63c0f2d9e717b4035c2926f52744ad29a70b0f5d84bae0a60d not found: ID does not exist" Nov 24 09:20:24 crc kubenswrapper[4944]: I1124 09:20:24.285760 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e28f479-229b-4d11-8d45-32b16edcc808" path="/var/lib/kubelet/pods/0e28f479-229b-4d11-8d45-32b16edcc808/volumes" Nov 24 09:20:26 crc kubenswrapper[4944]: I1124 09:20:26.277069 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:20:26 crc kubenswrapper[4944]: E1124 09:20:26.278018 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:20:28 crc kubenswrapper[4944]: I1124 09:20:28.130519 4944 scope.go:117] "RemoveContainer" containerID="41a16bdf6b3d1976b1e1aad714788e4ff1661da8136d3a26d52035543b68961c" Nov 24 09:20:28 crc kubenswrapper[4944]: I1124 09:20:28.155567 4944 scope.go:117] "RemoveContainer" containerID="baeed649469760ea552380de706d9f29c4df36ac76ccb2186b7193cf50ccbe23" Nov 24 09:20:28 crc kubenswrapper[4944]: I1124 09:20:28.200954 4944 scope.go:117] "RemoveContainer" containerID="873fb4ad2a63dd371217f9b304da3a326d22fb85587ad6f5c65d2bd833dca7af" Nov 24 09:20:28 crc kubenswrapper[4944]: I1124 09:20:28.220257 4944 scope.go:117] "RemoveContainer" containerID="5e472b6966fd4a5f6d4239a02c06bce03ae0ec88a39cd86b701265cd0f16922a" Nov 24 09:20:28 crc kubenswrapper[4944]: I1124 09:20:28.246692 4944 scope.go:117] "RemoveContainer" containerID="e427978bf154c1166e0f3fa4b8da09e0203655a4a66eff43c601c65fe8459316" Nov 24 09:20:28 crc kubenswrapper[4944]: I1124 09:20:28.306158 4944 scope.go:117] "RemoveContainer" containerID="87bc89b8a47e19382a1d4524937f8b6a5ab31d4623f694f16b7b41a3fda03f50" Nov 24 09:20:28 crc kubenswrapper[4944]: I1124 09:20:28.321754 4944 scope.go:117] "RemoveContainer" containerID="f79b603137a9e552f12da861f283799163e55db755214ecf4ce6c275a2f775c7" Nov 24 09:20:40 crc kubenswrapper[4944]: I1124 09:20:40.281292 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:20:40 crc kubenswrapper[4944]: E1124 09:20:40.283084 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:20:53 crc kubenswrapper[4944]: I1124 09:20:53.276482 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:20:53 crc kubenswrapper[4944]: E1124 09:20:53.278524 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:21:05 crc kubenswrapper[4944]: I1124 09:21:05.277037 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:21:05 crc kubenswrapper[4944]: E1124 09:21:05.277847 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:21:17 crc kubenswrapper[4944]: I1124 09:21:17.277903 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:21:17 crc kubenswrapper[4944]: E1124 09:21:17.282316 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:21:31 crc kubenswrapper[4944]: I1124 09:21:31.276900 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:21:31 crc kubenswrapper[4944]: E1124 09:21:31.277834 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:21:43 crc kubenswrapper[4944]: I1124 09:21:43.277224 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:21:43 crc kubenswrapper[4944]: E1124 09:21:43.277911 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:21:57 crc kubenswrapper[4944]: I1124 09:21:57.276468 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:21:57 crc kubenswrapper[4944]: E1124 09:21:57.277239 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:22:08 crc kubenswrapper[4944]: I1124 09:22:08.276951 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:22:08 crc kubenswrapper[4944]: E1124 09:22:08.277764 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:22:23 crc kubenswrapper[4944]: I1124 09:22:23.277352 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:22:23 crc kubenswrapper[4944]: E1124 09:22:23.278153 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:22:35 crc kubenswrapper[4944]: I1124 09:22:35.277347 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:22:35 crc kubenswrapper[4944]: E1124 09:22:35.279827 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:22:49 crc kubenswrapper[4944]: I1124 09:22:49.276796 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:22:49 crc kubenswrapper[4944]: E1124 09:22:49.277741 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:23:01 crc kubenswrapper[4944]: I1124 09:23:01.277038 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:23:01 crc kubenswrapper[4944]: I1124 09:23:01.426733 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"21a679bbbfed343720da0b5686133502e7373b5c23fd320c6e71026dbf0131e1"} Nov 24 09:25:23 crc kubenswrapper[4944]: I1124 09:25:23.548397 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:25:23 crc kubenswrapper[4944]: I1124 09:25:23.549227 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:25:53 crc kubenswrapper[4944]: I1124 09:25:53.549335 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:25:53 crc kubenswrapper[4944]: I1124 09:25:53.550222 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:26:12 crc kubenswrapper[4944]: I1124 09:26:12.898795 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m9gnw"] Nov 24 09:26:12 crc kubenswrapper[4944]: E1124 09:26:12.902293 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e28f479-229b-4d11-8d45-32b16edcc808" containerName="registry-server" Nov 24 09:26:12 crc kubenswrapper[4944]: I1124 09:26:12.902316 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e28f479-229b-4d11-8d45-32b16edcc808" containerName="registry-server" Nov 24 09:26:12 crc kubenswrapper[4944]: E1124 09:26:12.902333 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e28f479-229b-4d11-8d45-32b16edcc808" containerName="extract-content" Nov 24 09:26:12 crc kubenswrapper[4944]: I1124 09:26:12.902341 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e28f479-229b-4d11-8d45-32b16edcc808" containerName="extract-content" Nov 24 09:26:12 crc kubenswrapper[4944]: E1124 09:26:12.902372 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e28f479-229b-4d11-8d45-32b16edcc808" containerName="extract-utilities" Nov 24 09:26:12 crc kubenswrapper[4944]: I1124 09:26:12.902380 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e28f479-229b-4d11-8d45-32b16edcc808" containerName="extract-utilities" Nov 24 09:26:12 crc kubenswrapper[4944]: I1124 09:26:12.902734 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e28f479-229b-4d11-8d45-32b16edcc808" containerName="registry-server" Nov 24 09:26:12 crc kubenswrapper[4944]: I1124 09:26:12.904144 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:12 crc kubenswrapper[4944]: I1124 09:26:12.909528 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m9gnw"] Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.007060 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-utilities\") pod \"certified-operators-m9gnw\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.007134 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-catalog-content\") pod \"certified-operators-m9gnw\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.007192 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bcck\" (UniqueName: \"kubernetes.io/projected/d8f53458-0808-4af8-b408-87cceb3c7f4c-kube-api-access-7bcck\") pod \"certified-operators-m9gnw\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.108419 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-catalog-content\") pod \"certified-operators-m9gnw\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.108482 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bcck\" (UniqueName: \"kubernetes.io/projected/d8f53458-0808-4af8-b408-87cceb3c7f4c-kube-api-access-7bcck\") pod \"certified-operators-m9gnw\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.108567 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-utilities\") pod \"certified-operators-m9gnw\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.109039 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-utilities\") pod \"certified-operators-m9gnw\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.109073 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-catalog-content\") pod \"certified-operators-m9gnw\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.129481 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bcck\" (UniqueName: \"kubernetes.io/projected/d8f53458-0808-4af8-b408-87cceb3c7f4c-kube-api-access-7bcck\") pod \"certified-operators-m9gnw\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.236041 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.713357 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m9gnw"] Nov 24 09:26:13 crc kubenswrapper[4944]: I1124 09:26:13.785605 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m9gnw" event={"ID":"d8f53458-0808-4af8-b408-87cceb3c7f4c","Type":"ContainerStarted","Data":"d25f7cb810adb73ff06dec04f9f90c835dafda16b8a80bfb15a16bb6a3601909"} Nov 24 09:26:14 crc kubenswrapper[4944]: I1124 09:26:14.793474 4944 generic.go:334] "Generic (PLEG): container finished" podID="d8f53458-0808-4af8-b408-87cceb3c7f4c" containerID="06616cdf12d9a67597bcc76fddf04d9ad95fdb8f660bd852a08e94430825f623" exitCode=0 Nov 24 09:26:14 crc kubenswrapper[4944]: I1124 09:26:14.793584 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m9gnw" event={"ID":"d8f53458-0808-4af8-b408-87cceb3c7f4c","Type":"ContainerDied","Data":"06616cdf12d9a67597bcc76fddf04d9ad95fdb8f660bd852a08e94430825f623"} Nov 24 09:26:14 crc kubenswrapper[4944]: I1124 09:26:14.795400 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:26:15 crc kubenswrapper[4944]: I1124 09:26:15.803166 4944 generic.go:334] "Generic (PLEG): container finished" podID="d8f53458-0808-4af8-b408-87cceb3c7f4c" containerID="aeb5a1c8dd601606615e4804ce999e553eec093608d5ce73389a1ea2a0e7cc7b" exitCode=0 Nov 24 09:26:15 crc kubenswrapper[4944]: I1124 09:26:15.803207 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m9gnw" event={"ID":"d8f53458-0808-4af8-b408-87cceb3c7f4c","Type":"ContainerDied","Data":"aeb5a1c8dd601606615e4804ce999e553eec093608d5ce73389a1ea2a0e7cc7b"} Nov 24 09:26:16 crc kubenswrapper[4944]: I1124 09:26:16.817850 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m9gnw" event={"ID":"d8f53458-0808-4af8-b408-87cceb3c7f4c","Type":"ContainerStarted","Data":"4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887"} Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.236891 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.237217 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.281899 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.305137 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m9gnw" podStartSLOduration=9.753469767 podStartE2EDuration="11.305115034s" podCreationTimestamp="2025-11-24 09:26:12 +0000 UTC" firstStartedPulling="2025-11-24 09:26:14.795186227 +0000 UTC m=+2035.329626689" lastFinishedPulling="2025-11-24 09:26:16.346831494 +0000 UTC m=+2036.881271956" observedRunningTime="2025-11-24 09:26:16.835353094 +0000 UTC m=+2037.369793556" watchObservedRunningTime="2025-11-24 09:26:23.305115034 +0000 UTC m=+2043.839555496" Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.549023 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.549397 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.549568 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.550295 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"21a679bbbfed343720da0b5686133502e7373b5c23fd320c6e71026dbf0131e1"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.550441 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://21a679bbbfed343720da0b5686133502e7373b5c23fd320c6e71026dbf0131e1" gracePeriod=600 Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.869232 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="21a679bbbfed343720da0b5686133502e7373b5c23fd320c6e71026dbf0131e1" exitCode=0 Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.869320 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"21a679bbbfed343720da0b5686133502e7373b5c23fd320c6e71026dbf0131e1"} Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.869693 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7"} Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.869719 4944 scope.go:117] "RemoveContainer" containerID="43ec448a6ea3bfdfc2bc3373bde6c0b62bb629322112ddc641450157749cf731" Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.916446 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:23 crc kubenswrapper[4944]: I1124 09:26:23.957346 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m9gnw"] Nov 24 09:26:25 crc kubenswrapper[4944]: I1124 09:26:25.885431 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m9gnw" podUID="d8f53458-0808-4af8-b408-87cceb3c7f4c" containerName="registry-server" containerID="cri-o://4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887" gracePeriod=2 Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.756684 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.894007 4944 generic.go:334] "Generic (PLEG): container finished" podID="d8f53458-0808-4af8-b408-87cceb3c7f4c" containerID="4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887" exitCode=0 Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.894041 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m9gnw" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.894079 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m9gnw" event={"ID":"d8f53458-0808-4af8-b408-87cceb3c7f4c","Type":"ContainerDied","Data":"4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887"} Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.894431 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m9gnw" event={"ID":"d8f53458-0808-4af8-b408-87cceb3c7f4c","Type":"ContainerDied","Data":"d25f7cb810adb73ff06dec04f9f90c835dafda16b8a80bfb15a16bb6a3601909"} Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.894448 4944 scope.go:117] "RemoveContainer" containerID="4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.912445 4944 scope.go:117] "RemoveContainer" containerID="aeb5a1c8dd601606615e4804ce999e553eec093608d5ce73389a1ea2a0e7cc7b" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.937109 4944 scope.go:117] "RemoveContainer" containerID="06616cdf12d9a67597bcc76fddf04d9ad95fdb8f660bd852a08e94430825f623" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.940546 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-catalog-content\") pod \"d8f53458-0808-4af8-b408-87cceb3c7f4c\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.940658 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bcck\" (UniqueName: \"kubernetes.io/projected/d8f53458-0808-4af8-b408-87cceb3c7f4c-kube-api-access-7bcck\") pod \"d8f53458-0808-4af8-b408-87cceb3c7f4c\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.940726 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-utilities\") pod \"d8f53458-0808-4af8-b408-87cceb3c7f4c\" (UID: \"d8f53458-0808-4af8-b408-87cceb3c7f4c\") " Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.941792 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-utilities" (OuterVolumeSpecName: "utilities") pod "d8f53458-0808-4af8-b408-87cceb3c7f4c" (UID: "d8f53458-0808-4af8-b408-87cceb3c7f4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.945687 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f53458-0808-4af8-b408-87cceb3c7f4c-kube-api-access-7bcck" (OuterVolumeSpecName: "kube-api-access-7bcck") pod "d8f53458-0808-4af8-b408-87cceb3c7f4c" (UID: "d8f53458-0808-4af8-b408-87cceb3c7f4c"). InnerVolumeSpecName "kube-api-access-7bcck". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.984260 4944 scope.go:117] "RemoveContainer" containerID="4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887" Nov 24 09:26:26 crc kubenswrapper[4944]: E1124 09:26:26.984787 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887\": container with ID starting with 4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887 not found: ID does not exist" containerID="4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.984814 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887"} err="failed to get container status \"4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887\": rpc error: code = NotFound desc = could not find container \"4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887\": container with ID starting with 4cfc79deba19e858da1e40f81bdd6d0e59efaa6a50055233de7618f2d3ba5887 not found: ID does not exist" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.984835 4944 scope.go:117] "RemoveContainer" containerID="aeb5a1c8dd601606615e4804ce999e553eec093608d5ce73389a1ea2a0e7cc7b" Nov 24 09:26:26 crc kubenswrapper[4944]: E1124 09:26:26.985216 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aeb5a1c8dd601606615e4804ce999e553eec093608d5ce73389a1ea2a0e7cc7b\": container with ID starting with aeb5a1c8dd601606615e4804ce999e553eec093608d5ce73389a1ea2a0e7cc7b not found: ID does not exist" containerID="aeb5a1c8dd601606615e4804ce999e553eec093608d5ce73389a1ea2a0e7cc7b" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.985238 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeb5a1c8dd601606615e4804ce999e553eec093608d5ce73389a1ea2a0e7cc7b"} err="failed to get container status \"aeb5a1c8dd601606615e4804ce999e553eec093608d5ce73389a1ea2a0e7cc7b\": rpc error: code = NotFound desc = could not find container \"aeb5a1c8dd601606615e4804ce999e553eec093608d5ce73389a1ea2a0e7cc7b\": container with ID starting with aeb5a1c8dd601606615e4804ce999e553eec093608d5ce73389a1ea2a0e7cc7b not found: ID does not exist" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.985255 4944 scope.go:117] "RemoveContainer" containerID="06616cdf12d9a67597bcc76fddf04d9ad95fdb8f660bd852a08e94430825f623" Nov 24 09:26:26 crc kubenswrapper[4944]: E1124 09:26:26.985509 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06616cdf12d9a67597bcc76fddf04d9ad95fdb8f660bd852a08e94430825f623\": container with ID starting with 06616cdf12d9a67597bcc76fddf04d9ad95fdb8f660bd852a08e94430825f623 not found: ID does not exist" containerID="06616cdf12d9a67597bcc76fddf04d9ad95fdb8f660bd852a08e94430825f623" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.985536 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06616cdf12d9a67597bcc76fddf04d9ad95fdb8f660bd852a08e94430825f623"} err="failed to get container status \"06616cdf12d9a67597bcc76fddf04d9ad95fdb8f660bd852a08e94430825f623\": rpc error: code = NotFound desc = could not find container \"06616cdf12d9a67597bcc76fddf04d9ad95fdb8f660bd852a08e94430825f623\": container with ID starting with 06616cdf12d9a67597bcc76fddf04d9ad95fdb8f660bd852a08e94430825f623 not found: ID does not exist" Nov 24 09:26:26 crc kubenswrapper[4944]: I1124 09:26:26.987836 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8f53458-0808-4af8-b408-87cceb3c7f4c" (UID: "d8f53458-0808-4af8-b408-87cceb3c7f4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:26:27 crc kubenswrapper[4944]: I1124 09:26:27.042570 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:26:27 crc kubenswrapper[4944]: I1124 09:26:27.042610 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bcck\" (UniqueName: \"kubernetes.io/projected/d8f53458-0808-4af8-b408-87cceb3c7f4c-kube-api-access-7bcck\") on node \"crc\" DevicePath \"\"" Nov 24 09:26:27 crc kubenswrapper[4944]: I1124 09:26:27.042622 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8f53458-0808-4af8-b408-87cceb3c7f4c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:26:27 crc kubenswrapper[4944]: I1124 09:26:27.223599 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m9gnw"] Nov 24 09:26:27 crc kubenswrapper[4944]: I1124 09:26:27.228458 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m9gnw"] Nov 24 09:26:28 crc kubenswrapper[4944]: I1124 09:26:28.284980 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8f53458-0808-4af8-b408-87cceb3c7f4c" path="/var/lib/kubelet/pods/d8f53458-0808-4af8-b408-87cceb3c7f4c/volumes" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.212311 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cc9qg"] Nov 24 09:26:41 crc kubenswrapper[4944]: E1124 09:26:41.213257 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f53458-0808-4af8-b408-87cceb3c7f4c" containerName="extract-utilities" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.213275 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f53458-0808-4af8-b408-87cceb3c7f4c" containerName="extract-utilities" Nov 24 09:26:41 crc kubenswrapper[4944]: E1124 09:26:41.213292 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f53458-0808-4af8-b408-87cceb3c7f4c" containerName="registry-server" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.213300 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f53458-0808-4af8-b408-87cceb3c7f4c" containerName="registry-server" Nov 24 09:26:41 crc kubenswrapper[4944]: E1124 09:26:41.213319 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f53458-0808-4af8-b408-87cceb3c7f4c" containerName="extract-content" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.213327 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f53458-0808-4af8-b408-87cceb3c7f4c" containerName="extract-content" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.213500 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f53458-0808-4af8-b408-87cceb3c7f4c" containerName="registry-server" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.214680 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.228398 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cc9qg"] Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.299914 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-catalog-content\") pod \"redhat-operators-cc9qg\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.300644 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njmmn\" (UniqueName: \"kubernetes.io/projected/949c728c-a453-4320-b8fb-e08ca1048366-kube-api-access-njmmn\") pod \"redhat-operators-cc9qg\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.301404 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-utilities\") pod \"redhat-operators-cc9qg\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.403319 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-utilities\") pod \"redhat-operators-cc9qg\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.403777 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-utilities\") pod \"redhat-operators-cc9qg\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.403919 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-catalog-content\") pod \"redhat-operators-cc9qg\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.403943 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njmmn\" (UniqueName: \"kubernetes.io/projected/949c728c-a453-4320-b8fb-e08ca1048366-kube-api-access-njmmn\") pod \"redhat-operators-cc9qg\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.404385 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-catalog-content\") pod \"redhat-operators-cc9qg\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.424070 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njmmn\" (UniqueName: \"kubernetes.io/projected/949c728c-a453-4320-b8fb-e08ca1048366-kube-api-access-njmmn\") pod \"redhat-operators-cc9qg\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.532965 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.975345 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cc9qg"] Nov 24 09:26:41 crc kubenswrapper[4944]: I1124 09:26:41.989434 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cc9qg" event={"ID":"949c728c-a453-4320-b8fb-e08ca1048366","Type":"ContainerStarted","Data":"4c968e3f7d7e19eec5e323ac71a4b96e374d847e1145124a1005b8c4ff5b5652"} Nov 24 09:26:43 crc kubenswrapper[4944]: I1124 09:26:43.010226 4944 generic.go:334] "Generic (PLEG): container finished" podID="949c728c-a453-4320-b8fb-e08ca1048366" containerID="7792726009e25a542377ec9b77d465acccbe392eae92f159bc3d8c2c3959e694" exitCode=0 Nov 24 09:26:43 crc kubenswrapper[4944]: I1124 09:26:43.011481 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cc9qg" event={"ID":"949c728c-a453-4320-b8fb-e08ca1048366","Type":"ContainerDied","Data":"7792726009e25a542377ec9b77d465acccbe392eae92f159bc3d8c2c3959e694"} Nov 24 09:26:44 crc kubenswrapper[4944]: I1124 09:26:44.023779 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cc9qg" event={"ID":"949c728c-a453-4320-b8fb-e08ca1048366","Type":"ContainerStarted","Data":"942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f"} Nov 24 09:26:45 crc kubenswrapper[4944]: I1124 09:26:45.036514 4944 generic.go:334] "Generic (PLEG): container finished" podID="949c728c-a453-4320-b8fb-e08ca1048366" containerID="942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f" exitCode=0 Nov 24 09:26:45 crc kubenswrapper[4944]: I1124 09:26:45.036553 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cc9qg" event={"ID":"949c728c-a453-4320-b8fb-e08ca1048366","Type":"ContainerDied","Data":"942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f"} Nov 24 09:26:46 crc kubenswrapper[4944]: I1124 09:26:46.044637 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cc9qg" event={"ID":"949c728c-a453-4320-b8fb-e08ca1048366","Type":"ContainerStarted","Data":"8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10"} Nov 24 09:26:46 crc kubenswrapper[4944]: I1124 09:26:46.068234 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cc9qg" podStartSLOduration=2.519396011 podStartE2EDuration="5.0682147s" podCreationTimestamp="2025-11-24 09:26:41 +0000 UTC" firstStartedPulling="2025-11-24 09:26:43.012701491 +0000 UTC m=+2063.547141953" lastFinishedPulling="2025-11-24 09:26:45.56152018 +0000 UTC m=+2066.095960642" observedRunningTime="2025-11-24 09:26:46.063285053 +0000 UTC m=+2066.597725515" watchObservedRunningTime="2025-11-24 09:26:46.0682147 +0000 UTC m=+2066.602655162" Nov 24 09:26:51 crc kubenswrapper[4944]: I1124 09:26:51.534208 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:51 crc kubenswrapper[4944]: I1124 09:26:51.534513 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:51 crc kubenswrapper[4944]: I1124 09:26:51.573827 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:52 crc kubenswrapper[4944]: I1124 09:26:52.124236 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:52 crc kubenswrapper[4944]: I1124 09:26:52.167019 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cc9qg"] Nov 24 09:26:54 crc kubenswrapper[4944]: I1124 09:26:54.101888 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cc9qg" podUID="949c728c-a453-4320-b8fb-e08ca1048366" containerName="registry-server" containerID="cri-o://8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10" gracePeriod=2 Nov 24 09:26:55 crc kubenswrapper[4944]: I1124 09:26:55.665286 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:55 crc kubenswrapper[4944]: I1124 09:26:55.694277 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-catalog-content\") pod \"949c728c-a453-4320-b8fb-e08ca1048366\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " Nov 24 09:26:55 crc kubenswrapper[4944]: I1124 09:26:55.796216 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njmmn\" (UniqueName: \"kubernetes.io/projected/949c728c-a453-4320-b8fb-e08ca1048366-kube-api-access-njmmn\") pod \"949c728c-a453-4320-b8fb-e08ca1048366\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " Nov 24 09:26:55 crc kubenswrapper[4944]: I1124 09:26:55.796798 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-utilities\") pod \"949c728c-a453-4320-b8fb-e08ca1048366\" (UID: \"949c728c-a453-4320-b8fb-e08ca1048366\") " Nov 24 09:26:55 crc kubenswrapper[4944]: I1124 09:26:55.797770 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "949c728c-a453-4320-b8fb-e08ca1048366" (UID: "949c728c-a453-4320-b8fb-e08ca1048366"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:26:55 crc kubenswrapper[4944]: I1124 09:26:55.798091 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-utilities" (OuterVolumeSpecName: "utilities") pod "949c728c-a453-4320-b8fb-e08ca1048366" (UID: "949c728c-a453-4320-b8fb-e08ca1048366"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:26:55 crc kubenswrapper[4944]: I1124 09:26:55.802201 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/949c728c-a453-4320-b8fb-e08ca1048366-kube-api-access-njmmn" (OuterVolumeSpecName: "kube-api-access-njmmn") pod "949c728c-a453-4320-b8fb-e08ca1048366" (UID: "949c728c-a453-4320-b8fb-e08ca1048366"). InnerVolumeSpecName "kube-api-access-njmmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:26:55 crc kubenswrapper[4944]: I1124 09:26:55.897753 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njmmn\" (UniqueName: \"kubernetes.io/projected/949c728c-a453-4320-b8fb-e08ca1048366-kube-api-access-njmmn\") on node \"crc\" DevicePath \"\"" Nov 24 09:26:55 crc kubenswrapper[4944]: I1124 09:26:55.897802 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:26:55 crc kubenswrapper[4944]: I1124 09:26:55.897820 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/949c728c-a453-4320-b8fb-e08ca1048366-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.125691 4944 generic.go:334] "Generic (PLEG): container finished" podID="949c728c-a453-4320-b8fb-e08ca1048366" containerID="8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10" exitCode=0 Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.125741 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cc9qg" event={"ID":"949c728c-a453-4320-b8fb-e08ca1048366","Type":"ContainerDied","Data":"8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10"} Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.125770 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cc9qg" event={"ID":"949c728c-a453-4320-b8fb-e08ca1048366","Type":"ContainerDied","Data":"4c968e3f7d7e19eec5e323ac71a4b96e374d847e1145124a1005b8c4ff5b5652"} Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.125787 4944 scope.go:117] "RemoveContainer" containerID="8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10" Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.125801 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cc9qg" Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.146569 4944 scope.go:117] "RemoveContainer" containerID="942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f" Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.168304 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cc9qg"] Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.178913 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cc9qg"] Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.179805 4944 scope.go:117] "RemoveContainer" containerID="7792726009e25a542377ec9b77d465acccbe392eae92f159bc3d8c2c3959e694" Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.207347 4944 scope.go:117] "RemoveContainer" containerID="8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10" Nov 24 09:26:56 crc kubenswrapper[4944]: E1124 09:26:56.209581 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10\": container with ID starting with 8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10 not found: ID does not exist" containerID="8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10" Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.209621 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10"} err="failed to get container status \"8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10\": rpc error: code = NotFound desc = could not find container \"8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10\": container with ID starting with 8dd6b2a5cdef1447c4e10eb3dc7992b48efbccc2ae71885160975bb2f1930d10 not found: ID does not exist" Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.209649 4944 scope.go:117] "RemoveContainer" containerID="942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f" Nov 24 09:26:56 crc kubenswrapper[4944]: E1124 09:26:56.209897 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f\": container with ID starting with 942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f not found: ID does not exist" containerID="942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f" Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.209917 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f"} err="failed to get container status \"942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f\": rpc error: code = NotFound desc = could not find container \"942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f\": container with ID starting with 942326c2aedf7b55797ee4d387c67487f5523020f3f54afc3b1aa63e4e4e829f not found: ID does not exist" Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.209935 4944 scope.go:117] "RemoveContainer" containerID="7792726009e25a542377ec9b77d465acccbe392eae92f159bc3d8c2c3959e694" Nov 24 09:26:56 crc kubenswrapper[4944]: E1124 09:26:56.210216 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7792726009e25a542377ec9b77d465acccbe392eae92f159bc3d8c2c3959e694\": container with ID starting with 7792726009e25a542377ec9b77d465acccbe392eae92f159bc3d8c2c3959e694 not found: ID does not exist" containerID="7792726009e25a542377ec9b77d465acccbe392eae92f159bc3d8c2c3959e694" Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.210235 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7792726009e25a542377ec9b77d465acccbe392eae92f159bc3d8c2c3959e694"} err="failed to get container status \"7792726009e25a542377ec9b77d465acccbe392eae92f159bc3d8c2c3959e694\": rpc error: code = NotFound desc = could not find container \"7792726009e25a542377ec9b77d465acccbe392eae92f159bc3d8c2c3959e694\": container with ID starting with 7792726009e25a542377ec9b77d465acccbe392eae92f159bc3d8c2c3959e694 not found: ID does not exist" Nov 24 09:26:56 crc kubenswrapper[4944]: I1124 09:26:56.285155 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="949c728c-a453-4320-b8fb-e08ca1048366" path="/var/lib/kubelet/pods/949c728c-a453-4320-b8fb-e08ca1048366/volumes" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.563730 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5qrjz"] Nov 24 09:27:08 crc kubenswrapper[4944]: E1124 09:27:08.564749 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="949c728c-a453-4320-b8fb-e08ca1048366" containerName="extract-content" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.564768 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="949c728c-a453-4320-b8fb-e08ca1048366" containerName="extract-content" Nov 24 09:27:08 crc kubenswrapper[4944]: E1124 09:27:08.564785 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="949c728c-a453-4320-b8fb-e08ca1048366" containerName="extract-utilities" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.564795 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="949c728c-a453-4320-b8fb-e08ca1048366" containerName="extract-utilities" Nov 24 09:27:08 crc kubenswrapper[4944]: E1124 09:27:08.564819 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="949c728c-a453-4320-b8fb-e08ca1048366" containerName="registry-server" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.564829 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="949c728c-a453-4320-b8fb-e08ca1048366" containerName="registry-server" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.565098 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="949c728c-a453-4320-b8fb-e08ca1048366" containerName="registry-server" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.566758 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.586848 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qrjz"] Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.687907 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6t6d\" (UniqueName: \"kubernetes.io/projected/ba40ab61-7377-4b28-8052-770298705fd1-kube-api-access-t6t6d\") pod \"redhat-marketplace-5qrjz\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.687967 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-utilities\") pod \"redhat-marketplace-5qrjz\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.687985 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-catalog-content\") pod \"redhat-marketplace-5qrjz\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.789103 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6t6d\" (UniqueName: \"kubernetes.io/projected/ba40ab61-7377-4b28-8052-770298705fd1-kube-api-access-t6t6d\") pod \"redhat-marketplace-5qrjz\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.789180 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-utilities\") pod \"redhat-marketplace-5qrjz\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.789208 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-catalog-content\") pod \"redhat-marketplace-5qrjz\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.789751 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-utilities\") pod \"redhat-marketplace-5qrjz\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.789832 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-catalog-content\") pod \"redhat-marketplace-5qrjz\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.818659 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6t6d\" (UniqueName: \"kubernetes.io/projected/ba40ab61-7377-4b28-8052-770298705fd1-kube-api-access-t6t6d\") pod \"redhat-marketplace-5qrjz\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:08 crc kubenswrapper[4944]: I1124 09:27:08.884220 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:09 crc kubenswrapper[4944]: I1124 09:27:09.332965 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qrjz"] Nov 24 09:27:10 crc kubenswrapper[4944]: I1124 09:27:10.226018 4944 generic.go:334] "Generic (PLEG): container finished" podID="ba40ab61-7377-4b28-8052-770298705fd1" containerID="e3aa207f4e351736f6e1ca095cd687a345fe0249178f212b0d5d084ed1279d01" exitCode=0 Nov 24 09:27:10 crc kubenswrapper[4944]: I1124 09:27:10.226089 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qrjz" event={"ID":"ba40ab61-7377-4b28-8052-770298705fd1","Type":"ContainerDied","Data":"e3aa207f4e351736f6e1ca095cd687a345fe0249178f212b0d5d084ed1279d01"} Nov 24 09:27:10 crc kubenswrapper[4944]: I1124 09:27:10.226417 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qrjz" event={"ID":"ba40ab61-7377-4b28-8052-770298705fd1","Type":"ContainerStarted","Data":"c6b62c31b5e908e1328d57537851ff735fd4f9bc01cb57b1012fbd8efe38462c"} Nov 24 09:27:11 crc kubenswrapper[4944]: I1124 09:27:11.233718 4944 generic.go:334] "Generic (PLEG): container finished" podID="ba40ab61-7377-4b28-8052-770298705fd1" containerID="ebb832fbc40ee061f42e8a7dc42a6d1c88c8e54ba6b342fad585317ce7131b18" exitCode=0 Nov 24 09:27:11 crc kubenswrapper[4944]: I1124 09:27:11.233759 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qrjz" event={"ID":"ba40ab61-7377-4b28-8052-770298705fd1","Type":"ContainerDied","Data":"ebb832fbc40ee061f42e8a7dc42a6d1c88c8e54ba6b342fad585317ce7131b18"} Nov 24 09:27:12 crc kubenswrapper[4944]: I1124 09:27:12.257269 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qrjz" event={"ID":"ba40ab61-7377-4b28-8052-770298705fd1","Type":"ContainerStarted","Data":"932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6"} Nov 24 09:27:12 crc kubenswrapper[4944]: I1124 09:27:12.273634 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5qrjz" podStartSLOduration=2.838017612 podStartE2EDuration="4.273617888s" podCreationTimestamp="2025-11-24 09:27:08 +0000 UTC" firstStartedPulling="2025-11-24 09:27:10.228317737 +0000 UTC m=+2090.762758209" lastFinishedPulling="2025-11-24 09:27:11.663918023 +0000 UTC m=+2092.198358485" observedRunningTime="2025-11-24 09:27:12.272357347 +0000 UTC m=+2092.806797809" watchObservedRunningTime="2025-11-24 09:27:12.273617888 +0000 UTC m=+2092.808058350" Nov 24 09:27:18 crc kubenswrapper[4944]: I1124 09:27:18.884997 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:18 crc kubenswrapper[4944]: I1124 09:27:18.885617 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:18 crc kubenswrapper[4944]: I1124 09:27:18.924955 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:19 crc kubenswrapper[4944]: I1124 09:27:19.361456 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:19 crc kubenswrapper[4944]: I1124 09:27:19.400127 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qrjz"] Nov 24 09:27:21 crc kubenswrapper[4944]: I1124 09:27:21.316478 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5qrjz" podUID="ba40ab61-7377-4b28-8052-770298705fd1" containerName="registry-server" containerID="cri-o://932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6" gracePeriod=2 Nov 24 09:27:21 crc kubenswrapper[4944]: I1124 09:27:21.700931 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:21 crc kubenswrapper[4944]: I1124 09:27:21.814540 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6t6d\" (UniqueName: \"kubernetes.io/projected/ba40ab61-7377-4b28-8052-770298705fd1-kube-api-access-t6t6d\") pod \"ba40ab61-7377-4b28-8052-770298705fd1\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " Nov 24 09:27:21 crc kubenswrapper[4944]: I1124 09:27:21.814899 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-catalog-content\") pod \"ba40ab61-7377-4b28-8052-770298705fd1\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " Nov 24 09:27:21 crc kubenswrapper[4944]: I1124 09:27:21.815117 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-utilities\") pod \"ba40ab61-7377-4b28-8052-770298705fd1\" (UID: \"ba40ab61-7377-4b28-8052-770298705fd1\") " Nov 24 09:27:21 crc kubenswrapper[4944]: I1124 09:27:21.816650 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-utilities" (OuterVolumeSpecName: "utilities") pod "ba40ab61-7377-4b28-8052-770298705fd1" (UID: "ba40ab61-7377-4b28-8052-770298705fd1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:27:21 crc kubenswrapper[4944]: I1124 09:27:21.819987 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba40ab61-7377-4b28-8052-770298705fd1-kube-api-access-t6t6d" (OuterVolumeSpecName: "kube-api-access-t6t6d") pod "ba40ab61-7377-4b28-8052-770298705fd1" (UID: "ba40ab61-7377-4b28-8052-770298705fd1"). InnerVolumeSpecName "kube-api-access-t6t6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:27:21 crc kubenswrapper[4944]: I1124 09:27:21.832036 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba40ab61-7377-4b28-8052-770298705fd1" (UID: "ba40ab61-7377-4b28-8052-770298705fd1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:27:21 crc kubenswrapper[4944]: I1124 09:27:21.916493 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:27:21 crc kubenswrapper[4944]: I1124 09:27:21.916695 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba40ab61-7377-4b28-8052-770298705fd1-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:27:21 crc kubenswrapper[4944]: I1124 09:27:21.916778 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6t6d\" (UniqueName: \"kubernetes.io/projected/ba40ab61-7377-4b28-8052-770298705fd1-kube-api-access-t6t6d\") on node \"crc\" DevicePath \"\"" Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.323592 4944 generic.go:334] "Generic (PLEG): container finished" podID="ba40ab61-7377-4b28-8052-770298705fd1" containerID="932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6" exitCode=0 Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.323633 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qrjz" event={"ID":"ba40ab61-7377-4b28-8052-770298705fd1","Type":"ContainerDied","Data":"932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6"} Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.323662 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qrjz" event={"ID":"ba40ab61-7377-4b28-8052-770298705fd1","Type":"ContainerDied","Data":"c6b62c31b5e908e1328d57537851ff735fd4f9bc01cb57b1012fbd8efe38462c"} Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.323693 4944 scope.go:117] "RemoveContainer" containerID="932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6" Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.324201 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qrjz" Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.343783 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qrjz"] Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.346292 4944 scope.go:117] "RemoveContainer" containerID="ebb832fbc40ee061f42e8a7dc42a6d1c88c8e54ba6b342fad585317ce7131b18" Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.348503 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qrjz"] Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.367247 4944 scope.go:117] "RemoveContainer" containerID="e3aa207f4e351736f6e1ca095cd687a345fe0249178f212b0d5d084ed1279d01" Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.391032 4944 scope.go:117] "RemoveContainer" containerID="932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6" Nov 24 09:27:22 crc kubenswrapper[4944]: E1124 09:27:22.391601 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6\": container with ID starting with 932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6 not found: ID does not exist" containerID="932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6" Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.391653 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6"} err="failed to get container status \"932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6\": rpc error: code = NotFound desc = could not find container \"932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6\": container with ID starting with 932cbd6330271365a20438d07bc17c17066f1152b5d74eb7719ad156fba61ae6 not found: ID does not exist" Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.391685 4944 scope.go:117] "RemoveContainer" containerID="ebb832fbc40ee061f42e8a7dc42a6d1c88c8e54ba6b342fad585317ce7131b18" Nov 24 09:27:22 crc kubenswrapper[4944]: E1124 09:27:22.392016 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebb832fbc40ee061f42e8a7dc42a6d1c88c8e54ba6b342fad585317ce7131b18\": container with ID starting with ebb832fbc40ee061f42e8a7dc42a6d1c88c8e54ba6b342fad585317ce7131b18 not found: ID does not exist" containerID="ebb832fbc40ee061f42e8a7dc42a6d1c88c8e54ba6b342fad585317ce7131b18" Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.392059 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebb832fbc40ee061f42e8a7dc42a6d1c88c8e54ba6b342fad585317ce7131b18"} err="failed to get container status \"ebb832fbc40ee061f42e8a7dc42a6d1c88c8e54ba6b342fad585317ce7131b18\": rpc error: code = NotFound desc = could not find container \"ebb832fbc40ee061f42e8a7dc42a6d1c88c8e54ba6b342fad585317ce7131b18\": container with ID starting with ebb832fbc40ee061f42e8a7dc42a6d1c88c8e54ba6b342fad585317ce7131b18 not found: ID does not exist" Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.392074 4944 scope.go:117] "RemoveContainer" containerID="e3aa207f4e351736f6e1ca095cd687a345fe0249178f212b0d5d084ed1279d01" Nov 24 09:27:22 crc kubenswrapper[4944]: E1124 09:27:22.392282 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3aa207f4e351736f6e1ca095cd687a345fe0249178f212b0d5d084ed1279d01\": container with ID starting with e3aa207f4e351736f6e1ca095cd687a345fe0249178f212b0d5d084ed1279d01 not found: ID does not exist" containerID="e3aa207f4e351736f6e1ca095cd687a345fe0249178f212b0d5d084ed1279d01" Nov 24 09:27:22 crc kubenswrapper[4944]: I1124 09:27:22.392361 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3aa207f4e351736f6e1ca095cd687a345fe0249178f212b0d5d084ed1279d01"} err="failed to get container status \"e3aa207f4e351736f6e1ca095cd687a345fe0249178f212b0d5d084ed1279d01\": rpc error: code = NotFound desc = could not find container \"e3aa207f4e351736f6e1ca095cd687a345fe0249178f212b0d5d084ed1279d01\": container with ID starting with e3aa207f4e351736f6e1ca095cd687a345fe0249178f212b0d5d084ed1279d01 not found: ID does not exist" Nov 24 09:27:24 crc kubenswrapper[4944]: I1124 09:27:24.296696 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba40ab61-7377-4b28-8052-770298705fd1" path="/var/lib/kubelet/pods/ba40ab61-7377-4b28-8052-770298705fd1/volumes" Nov 24 09:28:23 crc kubenswrapper[4944]: I1124 09:28:23.548847 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:28:23 crc kubenswrapper[4944]: I1124 09:28:23.549621 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:28:53 crc kubenswrapper[4944]: I1124 09:28:53.548552 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:28:53 crc kubenswrapper[4944]: I1124 09:28:53.549207 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:29:23 crc kubenswrapper[4944]: I1124 09:29:23.548460 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:29:23 crc kubenswrapper[4944]: I1124 09:29:23.548897 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:29:23 crc kubenswrapper[4944]: I1124 09:29:23.548941 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:29:23 crc kubenswrapper[4944]: I1124 09:29:23.549519 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:29:23 crc kubenswrapper[4944]: I1124 09:29:23.549578 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" gracePeriod=600 Nov 24 09:29:23 crc kubenswrapper[4944]: E1124 09:29:23.675230 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:29:24 crc kubenswrapper[4944]: I1124 09:29:24.169801 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" exitCode=0 Nov 24 09:29:24 crc kubenswrapper[4944]: I1124 09:29:24.169850 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7"} Nov 24 09:29:24 crc kubenswrapper[4944]: I1124 09:29:24.169894 4944 scope.go:117] "RemoveContainer" containerID="21a679bbbfed343720da0b5686133502e7373b5c23fd320c6e71026dbf0131e1" Nov 24 09:29:24 crc kubenswrapper[4944]: I1124 09:29:24.170518 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:29:24 crc kubenswrapper[4944]: E1124 09:29:24.170931 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:29:34 crc kubenswrapper[4944]: I1124 09:29:34.277393 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:29:34 crc kubenswrapper[4944]: E1124 09:29:34.278125 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:29:45 crc kubenswrapper[4944]: I1124 09:29:45.277095 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:29:45 crc kubenswrapper[4944]: E1124 09:29:45.277831 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.146131 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz"] Nov 24 09:30:00 crc kubenswrapper[4944]: E1124 09:30:00.146993 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba40ab61-7377-4b28-8052-770298705fd1" containerName="extract-utilities" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.147011 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba40ab61-7377-4b28-8052-770298705fd1" containerName="extract-utilities" Nov 24 09:30:00 crc kubenswrapper[4944]: E1124 09:30:00.147103 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba40ab61-7377-4b28-8052-770298705fd1" containerName="registry-server" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.147115 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba40ab61-7377-4b28-8052-770298705fd1" containerName="registry-server" Nov 24 09:30:00 crc kubenswrapper[4944]: E1124 09:30:00.147127 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba40ab61-7377-4b28-8052-770298705fd1" containerName="extract-content" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.147134 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba40ab61-7377-4b28-8052-770298705fd1" containerName="extract-content" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.147324 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba40ab61-7377-4b28-8052-770298705fd1" containerName="registry-server" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.147865 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.153283 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.163543 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.166458 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz"] Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.271598 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xd5s\" (UniqueName: \"kubernetes.io/projected/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-kube-api-access-8xd5s\") pod \"collect-profiles-29399610-twmfz\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.271676 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-secret-volume\") pod \"collect-profiles-29399610-twmfz\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.271725 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-config-volume\") pod \"collect-profiles-29399610-twmfz\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.280435 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:30:00 crc kubenswrapper[4944]: E1124 09:30:00.280641 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.372506 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xd5s\" (UniqueName: \"kubernetes.io/projected/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-kube-api-access-8xd5s\") pod \"collect-profiles-29399610-twmfz\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.372576 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-secret-volume\") pod \"collect-profiles-29399610-twmfz\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.372618 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-config-volume\") pod \"collect-profiles-29399610-twmfz\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.373636 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-config-volume\") pod \"collect-profiles-29399610-twmfz\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.383906 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-secret-volume\") pod \"collect-profiles-29399610-twmfz\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.388143 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xd5s\" (UniqueName: \"kubernetes.io/projected/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-kube-api-access-8xd5s\") pod \"collect-profiles-29399610-twmfz\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.478442 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:00 crc kubenswrapper[4944]: I1124 09:30:00.745730 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz"] Nov 24 09:30:01 crc kubenswrapper[4944]: I1124 09:30:01.466016 4944 generic.go:334] "Generic (PLEG): container finished" podID="3f5f63a4-f6aa-4656-83ec-884f6c4f52bf" containerID="91035c422b3f43b0d85fe46b85fdbaaea59ef42de502b69aef1d1ef5d9d6eacd" exitCode=0 Nov 24 09:30:01 crc kubenswrapper[4944]: I1124 09:30:01.466087 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" event={"ID":"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf","Type":"ContainerDied","Data":"91035c422b3f43b0d85fe46b85fdbaaea59ef42de502b69aef1d1ef5d9d6eacd"} Nov 24 09:30:01 crc kubenswrapper[4944]: I1124 09:30:01.467507 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" event={"ID":"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf","Type":"ContainerStarted","Data":"6cc8ffdd1743caf3b3c4385f41ed5dedcb0e7ca521edd4ad406d7cedc09487ab"} Nov 24 09:30:02 crc kubenswrapper[4944]: I1124 09:30:02.870397 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.058070 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-secret-volume\") pod \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.058249 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xd5s\" (UniqueName: \"kubernetes.io/projected/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-kube-api-access-8xd5s\") pod \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.058330 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-config-volume\") pod \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\" (UID: \"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf\") " Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.058808 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-config-volume" (OuterVolumeSpecName: "config-volume") pod "3f5f63a4-f6aa-4656-83ec-884f6c4f52bf" (UID: "3f5f63a4-f6aa-4656-83ec-884f6c4f52bf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.063438 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3f5f63a4-f6aa-4656-83ec-884f6c4f52bf" (UID: "3f5f63a4-f6aa-4656-83ec-884f6c4f52bf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.063994 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-kube-api-access-8xd5s" (OuterVolumeSpecName: "kube-api-access-8xd5s") pod "3f5f63a4-f6aa-4656-83ec-884f6c4f52bf" (UID: "3f5f63a4-f6aa-4656-83ec-884f6c4f52bf"). InnerVolumeSpecName "kube-api-access-8xd5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.159866 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.159903 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.159917 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xd5s\" (UniqueName: \"kubernetes.io/projected/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf-kube-api-access-8xd5s\") on node \"crc\" DevicePath \"\"" Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.480307 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" event={"ID":"3f5f63a4-f6aa-4656-83ec-884f6c4f52bf","Type":"ContainerDied","Data":"6cc8ffdd1743caf3b3c4385f41ed5dedcb0e7ca521edd4ad406d7cedc09487ab"} Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.480352 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cc8ffdd1743caf3b3c4385f41ed5dedcb0e7ca521edd4ad406d7cedc09487ab" Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.480370 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz" Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.941818 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s"] Nov 24 09:30:03 crc kubenswrapper[4944]: I1124 09:30:03.949699 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-9zv4s"] Nov 24 09:30:04 crc kubenswrapper[4944]: I1124 09:30:04.286434 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0768e45f-6339-47f6-ba67-ffe401d2a502" path="/var/lib/kubelet/pods/0768e45f-6339-47f6-ba67-ffe401d2a502/volumes" Nov 24 09:30:15 crc kubenswrapper[4944]: I1124 09:30:15.277224 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:30:15 crc kubenswrapper[4944]: E1124 09:30:15.277901 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:30:28 crc kubenswrapper[4944]: I1124 09:30:28.747502 4944 scope.go:117] "RemoveContainer" containerID="f8c7fff3a079662f189454565ca21ea4eb2ffa8ce1ccecdd2eb63949e41cae04" Nov 24 09:30:29 crc kubenswrapper[4944]: I1124 09:30:29.277008 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:30:29 crc kubenswrapper[4944]: E1124 09:30:29.277277 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:30:35 crc kubenswrapper[4944]: I1124 09:30:35.881740 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p44l5"] Nov 24 09:30:35 crc kubenswrapper[4944]: E1124 09:30:35.882373 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f5f63a4-f6aa-4656-83ec-884f6c4f52bf" containerName="collect-profiles" Nov 24 09:30:35 crc kubenswrapper[4944]: I1124 09:30:35.882389 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f5f63a4-f6aa-4656-83ec-884f6c4f52bf" containerName="collect-profiles" Nov 24 09:30:35 crc kubenswrapper[4944]: I1124 09:30:35.882544 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f5f63a4-f6aa-4656-83ec-884f6c4f52bf" containerName="collect-profiles" Nov 24 09:30:35 crc kubenswrapper[4944]: I1124 09:30:35.883631 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:35 crc kubenswrapper[4944]: I1124 09:30:35.892325 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p44l5"] Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.037499 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-utilities\") pod \"community-operators-p44l5\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.037755 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-catalog-content\") pod \"community-operators-p44l5\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.037794 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l72r\" (UniqueName: \"kubernetes.io/projected/9770fb6a-7f60-45ee-b063-7c4255ebaf08-kube-api-access-7l72r\") pod \"community-operators-p44l5\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.139973 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-utilities\") pod \"community-operators-p44l5\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.140036 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-catalog-content\") pod \"community-operators-p44l5\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.140094 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l72r\" (UniqueName: \"kubernetes.io/projected/9770fb6a-7f60-45ee-b063-7c4255ebaf08-kube-api-access-7l72r\") pod \"community-operators-p44l5\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.141522 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-catalog-content\") pod \"community-operators-p44l5\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.141606 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-utilities\") pod \"community-operators-p44l5\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.163833 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l72r\" (UniqueName: \"kubernetes.io/projected/9770fb6a-7f60-45ee-b063-7c4255ebaf08-kube-api-access-7l72r\") pod \"community-operators-p44l5\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.217331 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.479785 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p44l5"] Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.749751 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p44l5" event={"ID":"9770fb6a-7f60-45ee-b063-7c4255ebaf08","Type":"ContainerStarted","Data":"286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c"} Nov 24 09:30:36 crc kubenswrapper[4944]: I1124 09:30:36.749813 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p44l5" event={"ID":"9770fb6a-7f60-45ee-b063-7c4255ebaf08","Type":"ContainerStarted","Data":"1d9a1792fd3440fdf9fd58aa8bf8bc5d9d9e8518f48ac2c0c8fa940b29a8f1bb"} Nov 24 09:30:37 crc kubenswrapper[4944]: I1124 09:30:37.762698 4944 generic.go:334] "Generic (PLEG): container finished" podID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" containerID="286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c" exitCode=0 Nov 24 09:30:37 crc kubenswrapper[4944]: I1124 09:30:37.762764 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p44l5" event={"ID":"9770fb6a-7f60-45ee-b063-7c4255ebaf08","Type":"ContainerDied","Data":"286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c"} Nov 24 09:30:39 crc kubenswrapper[4944]: I1124 09:30:39.780541 4944 generic.go:334] "Generic (PLEG): container finished" podID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" containerID="369eaf5fc72a2fc46b577813e3bc10180ff602b3f0a097fa9b71341af5e80874" exitCode=0 Nov 24 09:30:39 crc kubenswrapper[4944]: I1124 09:30:39.780650 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p44l5" event={"ID":"9770fb6a-7f60-45ee-b063-7c4255ebaf08","Type":"ContainerDied","Data":"369eaf5fc72a2fc46b577813e3bc10180ff602b3f0a097fa9b71341af5e80874"} Nov 24 09:30:41 crc kubenswrapper[4944]: I1124 09:30:41.795593 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p44l5" event={"ID":"9770fb6a-7f60-45ee-b063-7c4255ebaf08","Type":"ContainerStarted","Data":"94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6"} Nov 24 09:30:41 crc kubenswrapper[4944]: I1124 09:30:41.812925 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p44l5" podStartSLOduration=3.675668517 podStartE2EDuration="6.812905907s" podCreationTimestamp="2025-11-24 09:30:35 +0000 UTC" firstStartedPulling="2025-11-24 09:30:37.764253169 +0000 UTC m=+2298.298693641" lastFinishedPulling="2025-11-24 09:30:40.901490569 +0000 UTC m=+2301.435931031" observedRunningTime="2025-11-24 09:30:41.811853993 +0000 UTC m=+2302.346294505" watchObservedRunningTime="2025-11-24 09:30:41.812905907 +0000 UTC m=+2302.347346369" Nov 24 09:30:44 crc kubenswrapper[4944]: I1124 09:30:44.277195 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:30:44 crc kubenswrapper[4944]: E1124 09:30:44.277743 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:30:46 crc kubenswrapper[4944]: I1124 09:30:46.217559 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:46 crc kubenswrapper[4944]: I1124 09:30:46.217880 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:46 crc kubenswrapper[4944]: I1124 09:30:46.265022 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:46 crc kubenswrapper[4944]: I1124 09:30:46.884811 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:46 crc kubenswrapper[4944]: I1124 09:30:46.930547 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p44l5"] Nov 24 09:30:48 crc kubenswrapper[4944]: I1124 09:30:48.849321 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p44l5" podUID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" containerName="registry-server" containerID="cri-o://94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6" gracePeriod=2 Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.246810 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.330419 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7l72r\" (UniqueName: \"kubernetes.io/projected/9770fb6a-7f60-45ee-b063-7c4255ebaf08-kube-api-access-7l72r\") pod \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.330504 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-catalog-content\") pod \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.330565 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-utilities\") pod \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\" (UID: \"9770fb6a-7f60-45ee-b063-7c4255ebaf08\") " Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.331687 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-utilities" (OuterVolumeSpecName: "utilities") pod "9770fb6a-7f60-45ee-b063-7c4255ebaf08" (UID: "9770fb6a-7f60-45ee-b063-7c4255ebaf08"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.335557 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9770fb6a-7f60-45ee-b063-7c4255ebaf08-kube-api-access-7l72r" (OuterVolumeSpecName: "kube-api-access-7l72r") pod "9770fb6a-7f60-45ee-b063-7c4255ebaf08" (UID: "9770fb6a-7f60-45ee-b063-7c4255ebaf08"). InnerVolumeSpecName "kube-api-access-7l72r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.396301 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9770fb6a-7f60-45ee-b063-7c4255ebaf08" (UID: "9770fb6a-7f60-45ee-b063-7c4255ebaf08"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.432324 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.432356 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9770fb6a-7f60-45ee-b063-7c4255ebaf08-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.432367 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7l72r\" (UniqueName: \"kubernetes.io/projected/9770fb6a-7f60-45ee-b063-7c4255ebaf08-kube-api-access-7l72r\") on node \"crc\" DevicePath \"\"" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.856806 4944 generic.go:334] "Generic (PLEG): container finished" podID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" containerID="94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6" exitCode=0 Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.856858 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p44l5" event={"ID":"9770fb6a-7f60-45ee-b063-7c4255ebaf08","Type":"ContainerDied","Data":"94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6"} Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.857166 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p44l5" event={"ID":"9770fb6a-7f60-45ee-b063-7c4255ebaf08","Type":"ContainerDied","Data":"1d9a1792fd3440fdf9fd58aa8bf8bc5d9d9e8518f48ac2c0c8fa940b29a8f1bb"} Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.857189 4944 scope.go:117] "RemoveContainer" containerID="94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.856885 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p44l5" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.874851 4944 scope.go:117] "RemoveContainer" containerID="369eaf5fc72a2fc46b577813e3bc10180ff602b3f0a097fa9b71341af5e80874" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.884650 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p44l5"] Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.890496 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p44l5"] Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.914362 4944 scope.go:117] "RemoveContainer" containerID="286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.927672 4944 scope.go:117] "RemoveContainer" containerID="94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6" Nov 24 09:30:49 crc kubenswrapper[4944]: E1124 09:30:49.928018 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6\": container with ID starting with 94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6 not found: ID does not exist" containerID="94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.928084 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6"} err="failed to get container status \"94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6\": rpc error: code = NotFound desc = could not find container \"94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6\": container with ID starting with 94e412e7b35ac93abc2cb87fec2a6ff6579b53d670e4b14e1feeaf8d719c53c6 not found: ID does not exist" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.928120 4944 scope.go:117] "RemoveContainer" containerID="369eaf5fc72a2fc46b577813e3bc10180ff602b3f0a097fa9b71341af5e80874" Nov 24 09:30:49 crc kubenswrapper[4944]: E1124 09:30:49.928449 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"369eaf5fc72a2fc46b577813e3bc10180ff602b3f0a097fa9b71341af5e80874\": container with ID starting with 369eaf5fc72a2fc46b577813e3bc10180ff602b3f0a097fa9b71341af5e80874 not found: ID does not exist" containerID="369eaf5fc72a2fc46b577813e3bc10180ff602b3f0a097fa9b71341af5e80874" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.928485 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"369eaf5fc72a2fc46b577813e3bc10180ff602b3f0a097fa9b71341af5e80874"} err="failed to get container status \"369eaf5fc72a2fc46b577813e3bc10180ff602b3f0a097fa9b71341af5e80874\": rpc error: code = NotFound desc = could not find container \"369eaf5fc72a2fc46b577813e3bc10180ff602b3f0a097fa9b71341af5e80874\": container with ID starting with 369eaf5fc72a2fc46b577813e3bc10180ff602b3f0a097fa9b71341af5e80874 not found: ID does not exist" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.928513 4944 scope.go:117] "RemoveContainer" containerID="286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c" Nov 24 09:30:49 crc kubenswrapper[4944]: E1124 09:30:49.928806 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c\": container with ID starting with 286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c not found: ID does not exist" containerID="286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c" Nov 24 09:30:49 crc kubenswrapper[4944]: I1124 09:30:49.928830 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c"} err="failed to get container status \"286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c\": rpc error: code = NotFound desc = could not find container \"286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c\": container with ID starting with 286ac9caaa373e569468010ea3f6ac070b36115fd4b2f2e55a51250512ca599c not found: ID does not exist" Nov 24 09:30:50 crc kubenswrapper[4944]: I1124 09:30:50.284418 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" path="/var/lib/kubelet/pods/9770fb6a-7f60-45ee-b063-7c4255ebaf08/volumes" Nov 24 09:30:57 crc kubenswrapper[4944]: I1124 09:30:57.276217 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:30:57 crc kubenswrapper[4944]: E1124 09:30:57.276805 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:31:12 crc kubenswrapper[4944]: I1124 09:31:12.276576 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:31:12 crc kubenswrapper[4944]: E1124 09:31:12.277230 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:31:25 crc kubenswrapper[4944]: I1124 09:31:25.276712 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:31:25 crc kubenswrapper[4944]: E1124 09:31:25.277328 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:31:36 crc kubenswrapper[4944]: I1124 09:31:36.276528 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:31:36 crc kubenswrapper[4944]: E1124 09:31:36.277110 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:31:51 crc kubenswrapper[4944]: I1124 09:31:51.276374 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:31:51 crc kubenswrapper[4944]: E1124 09:31:51.277029 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:32:03 crc kubenswrapper[4944]: I1124 09:32:03.276492 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:32:03 crc kubenswrapper[4944]: E1124 09:32:03.277299 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:32:18 crc kubenswrapper[4944]: I1124 09:32:18.277418 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:32:18 crc kubenswrapper[4944]: E1124 09:32:18.279230 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:32:32 crc kubenswrapper[4944]: I1124 09:32:32.279808 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:32:32 crc kubenswrapper[4944]: E1124 09:32:32.280637 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:32:44 crc kubenswrapper[4944]: I1124 09:32:44.276996 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:32:44 crc kubenswrapper[4944]: E1124 09:32:44.277782 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:32:55 crc kubenswrapper[4944]: I1124 09:32:55.276701 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:32:55 crc kubenswrapper[4944]: E1124 09:32:55.277391 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:33:09 crc kubenswrapper[4944]: I1124 09:33:09.276712 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:33:09 crc kubenswrapper[4944]: E1124 09:33:09.277458 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:33:21 crc kubenswrapper[4944]: I1124 09:33:21.276836 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:33:21 crc kubenswrapper[4944]: E1124 09:33:21.277806 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:33:33 crc kubenswrapper[4944]: I1124 09:33:33.276789 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:33:33 crc kubenswrapper[4944]: E1124 09:33:33.277563 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:33:46 crc kubenswrapper[4944]: I1124 09:33:46.277207 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:33:46 crc kubenswrapper[4944]: E1124 09:33:46.277984 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:33:58 crc kubenswrapper[4944]: I1124 09:33:58.277661 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:33:58 crc kubenswrapper[4944]: E1124 09:33:58.278372 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:34:11 crc kubenswrapper[4944]: I1124 09:34:11.276975 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:34:11 crc kubenswrapper[4944]: E1124 09:34:11.277725 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:34:24 crc kubenswrapper[4944]: I1124 09:34:24.277227 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:34:25 crc kubenswrapper[4944]: I1124 09:34:25.354940 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"f7d69e51c92939bdcd5b3f695ca62927d0f1b11c6749722c6131860a7f19cb18"} Nov 24 09:36:47 crc kubenswrapper[4944]: I1124 09:36:47.940428 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5w2n6"] Nov 24 09:36:47 crc kubenswrapper[4944]: E1124 09:36:47.941527 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" containerName="extract-utilities" Nov 24 09:36:47 crc kubenswrapper[4944]: I1124 09:36:47.941547 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" containerName="extract-utilities" Nov 24 09:36:47 crc kubenswrapper[4944]: E1124 09:36:47.941575 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" containerName="registry-server" Nov 24 09:36:47 crc kubenswrapper[4944]: I1124 09:36:47.941584 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" containerName="registry-server" Nov 24 09:36:47 crc kubenswrapper[4944]: E1124 09:36:47.941598 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" containerName="extract-content" Nov 24 09:36:47 crc kubenswrapper[4944]: I1124 09:36:47.941606 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" containerName="extract-content" Nov 24 09:36:47 crc kubenswrapper[4944]: I1124 09:36:47.941795 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9770fb6a-7f60-45ee-b063-7c4255ebaf08" containerName="registry-server" Nov 24 09:36:47 crc kubenswrapper[4944]: I1124 09:36:47.943158 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:47 crc kubenswrapper[4944]: I1124 09:36:47.974933 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5w2n6"] Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.081747 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-catalog-content\") pod \"redhat-operators-5w2n6\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.081808 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-utilities\") pod \"redhat-operators-5w2n6\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.081869 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kp9x\" (UniqueName: \"kubernetes.io/projected/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-kube-api-access-6kp9x\") pod \"redhat-operators-5w2n6\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.183333 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-catalog-content\") pod \"redhat-operators-5w2n6\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.183401 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-utilities\") pod \"redhat-operators-5w2n6\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.183443 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kp9x\" (UniqueName: \"kubernetes.io/projected/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-kube-api-access-6kp9x\") pod \"redhat-operators-5w2n6\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.183989 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-catalog-content\") pod \"redhat-operators-5w2n6\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.184028 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-utilities\") pod \"redhat-operators-5w2n6\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.204020 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kp9x\" (UniqueName: \"kubernetes.io/projected/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-kube-api-access-6kp9x\") pod \"redhat-operators-5w2n6\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.270095 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.530357 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5w2n6"] Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.932488 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b4pv4"] Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.934270 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.948522 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b4pv4"] Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.996708 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlkgt\" (UniqueName: \"kubernetes.io/projected/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-kube-api-access-zlkgt\") pod \"certified-operators-b4pv4\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.996988 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-catalog-content\") pod \"certified-operators-b4pv4\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:48 crc kubenswrapper[4944]: I1124 09:36:48.997111 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-utilities\") pod \"certified-operators-b4pv4\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.048325 4944 generic.go:334] "Generic (PLEG): container finished" podID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" containerID="8c83d7c2d3ba033f24b3fbde739bb2b3705727d3ce88e64acfcf7083bbff3bf0" exitCode=0 Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.048378 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w2n6" event={"ID":"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70","Type":"ContainerDied","Data":"8c83d7c2d3ba033f24b3fbde739bb2b3705727d3ce88e64acfcf7083bbff3bf0"} Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.048410 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w2n6" event={"ID":"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70","Type":"ContainerStarted","Data":"d4689c0f01a6c6119d76c1b13ea03e18a00df439bbfc42dbf3f83210259c1301"} Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.050150 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.098130 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-catalog-content\") pod \"certified-operators-b4pv4\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.098200 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-utilities\") pod \"certified-operators-b4pv4\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.098673 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlkgt\" (UniqueName: \"kubernetes.io/projected/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-kube-api-access-zlkgt\") pod \"certified-operators-b4pv4\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.098841 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-utilities\") pod \"certified-operators-b4pv4\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.098713 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-catalog-content\") pod \"certified-operators-b4pv4\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.119014 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlkgt\" (UniqueName: \"kubernetes.io/projected/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-kube-api-access-zlkgt\") pod \"certified-operators-b4pv4\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.252509 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:49 crc kubenswrapper[4944]: I1124 09:36:49.701694 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b4pv4"] Nov 24 09:36:49 crc kubenswrapper[4944]: W1124 09:36:49.707714 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cccc8c7_4f29_4d86_9463_e2c71d4aaafb.slice/crio-3a670fd63cf00f8634b101767e82dc650d852867b0206738918027c7ba66f19b WatchSource:0}: Error finding container 3a670fd63cf00f8634b101767e82dc650d852867b0206738918027c7ba66f19b: Status 404 returned error can't find the container with id 3a670fd63cf00f8634b101767e82dc650d852867b0206738918027c7ba66f19b Nov 24 09:36:50 crc kubenswrapper[4944]: I1124 09:36:50.064732 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w2n6" event={"ID":"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70","Type":"ContainerStarted","Data":"7b0cbfe239499374f05bf13664afb0db758d92bb13bcb9d38f5d01da5287e4c6"} Nov 24 09:36:50 crc kubenswrapper[4944]: I1124 09:36:50.067402 4944 generic.go:334] "Generic (PLEG): container finished" podID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" containerID="93152006e2f6bd4f123ad5111eee2a52778e9d7b9c8388c76bea0064e888499e" exitCode=0 Nov 24 09:36:50 crc kubenswrapper[4944]: I1124 09:36:50.067434 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4pv4" event={"ID":"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb","Type":"ContainerDied","Data":"93152006e2f6bd4f123ad5111eee2a52778e9d7b9c8388c76bea0064e888499e"} Nov 24 09:36:50 crc kubenswrapper[4944]: I1124 09:36:50.067479 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4pv4" event={"ID":"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb","Type":"ContainerStarted","Data":"3a670fd63cf00f8634b101767e82dc650d852867b0206738918027c7ba66f19b"} Nov 24 09:36:51 crc kubenswrapper[4944]: I1124 09:36:51.077636 4944 generic.go:334] "Generic (PLEG): container finished" podID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" containerID="7b0cbfe239499374f05bf13664afb0db758d92bb13bcb9d38f5d01da5287e4c6" exitCode=0 Nov 24 09:36:51 crc kubenswrapper[4944]: I1124 09:36:51.077715 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w2n6" event={"ID":"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70","Type":"ContainerDied","Data":"7b0cbfe239499374f05bf13664afb0db758d92bb13bcb9d38f5d01da5287e4c6"} Nov 24 09:36:51 crc kubenswrapper[4944]: I1124 09:36:51.084918 4944 generic.go:334] "Generic (PLEG): container finished" podID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" containerID="9bf66e89bf8acb34bc7f9008a01b553149578db4b06877113e1b5da6db5c1b76" exitCode=0 Nov 24 09:36:51 crc kubenswrapper[4944]: I1124 09:36:51.084965 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4pv4" event={"ID":"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb","Type":"ContainerDied","Data":"9bf66e89bf8acb34bc7f9008a01b553149578db4b06877113e1b5da6db5c1b76"} Nov 24 09:36:52 crc kubenswrapper[4944]: I1124 09:36:52.094059 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w2n6" event={"ID":"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70","Type":"ContainerStarted","Data":"071afdaeafc8c930018a27809e8a431ecd7ace12f0e6d931b63fc811d8fe5c5e"} Nov 24 09:36:52 crc kubenswrapper[4944]: I1124 09:36:52.096395 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4pv4" event={"ID":"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb","Type":"ContainerStarted","Data":"c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790"} Nov 24 09:36:52 crc kubenswrapper[4944]: I1124 09:36:52.114992 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5w2n6" podStartSLOduration=2.725341787 podStartE2EDuration="5.114974449s" podCreationTimestamp="2025-11-24 09:36:47 +0000 UTC" firstStartedPulling="2025-11-24 09:36:49.049880875 +0000 UTC m=+2669.584321327" lastFinishedPulling="2025-11-24 09:36:51.439513527 +0000 UTC m=+2671.973953989" observedRunningTime="2025-11-24 09:36:52.108783442 +0000 UTC m=+2672.643223904" watchObservedRunningTime="2025-11-24 09:36:52.114974449 +0000 UTC m=+2672.649414911" Nov 24 09:36:52 crc kubenswrapper[4944]: I1124 09:36:52.128488 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b4pv4" podStartSLOduration=2.739016617 podStartE2EDuration="4.128469849s" podCreationTimestamp="2025-11-24 09:36:48 +0000 UTC" firstStartedPulling="2025-11-24 09:36:50.072409987 +0000 UTC m=+2670.606850449" lastFinishedPulling="2025-11-24 09:36:51.461863219 +0000 UTC m=+2671.996303681" observedRunningTime="2025-11-24 09:36:52.124506854 +0000 UTC m=+2672.658947316" watchObservedRunningTime="2025-11-24 09:36:52.128469849 +0000 UTC m=+2672.662910311" Nov 24 09:36:53 crc kubenswrapper[4944]: I1124 09:36:53.548575 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:36:53 crc kubenswrapper[4944]: I1124 09:36:53.548645 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:36:58 crc kubenswrapper[4944]: I1124 09:36:58.270731 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:58 crc kubenswrapper[4944]: I1124 09:36:58.271333 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:58 crc kubenswrapper[4944]: I1124 09:36:58.321816 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:59 crc kubenswrapper[4944]: I1124 09:36:59.214907 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:36:59 crc kubenswrapper[4944]: I1124 09:36:59.253405 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:59 crc kubenswrapper[4944]: I1124 09:36:59.254435 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:36:59 crc kubenswrapper[4944]: I1124 09:36:59.263216 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5w2n6"] Nov 24 09:36:59 crc kubenswrapper[4944]: I1124 09:36:59.293926 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:37:00 crc kubenswrapper[4944]: I1124 09:37:00.221245 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:37:00 crc kubenswrapper[4944]: I1124 09:37:00.953641 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b4pv4"] Nov 24 09:37:01 crc kubenswrapper[4944]: I1124 09:37:01.190456 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5w2n6" podUID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" containerName="registry-server" containerID="cri-o://071afdaeafc8c930018a27809e8a431ecd7ace12f0e6d931b63fc811d8fe5c5e" gracePeriod=2 Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.205637 4944 generic.go:334] "Generic (PLEG): container finished" podID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" containerID="071afdaeafc8c930018a27809e8a431ecd7ace12f0e6d931b63fc811d8fe5c5e" exitCode=0 Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.205721 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w2n6" event={"ID":"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70","Type":"ContainerDied","Data":"071afdaeafc8c930018a27809e8a431ecd7ace12f0e6d931b63fc811d8fe5c5e"} Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.206097 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w2n6" event={"ID":"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70","Type":"ContainerDied","Data":"d4689c0f01a6c6119d76c1b13ea03e18a00df439bbfc42dbf3f83210259c1301"} Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.206113 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4689c0f01a6c6119d76c1b13ea03e18a00df439bbfc42dbf3f83210259c1301" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.206276 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b4pv4" podUID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" containerName="registry-server" containerID="cri-o://c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790" gracePeriod=2 Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.229964 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.415149 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kp9x\" (UniqueName: \"kubernetes.io/projected/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-kube-api-access-6kp9x\") pod \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.415214 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-utilities\") pod \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.415241 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-catalog-content\") pod \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\" (UID: \"8ae16d21-1bb0-47c2-9c55-66bfb00ebd70\") " Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.416239 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-utilities" (OuterVolumeSpecName: "utilities") pod "8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" (UID: "8ae16d21-1bb0-47c2-9c55-66bfb00ebd70"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.426327 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-kube-api-access-6kp9x" (OuterVolumeSpecName: "kube-api-access-6kp9x") pod "8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" (UID: "8ae16d21-1bb0-47c2-9c55-66bfb00ebd70"). InnerVolumeSpecName "kube-api-access-6kp9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.516411 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kp9x\" (UniqueName: \"kubernetes.io/projected/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-kube-api-access-6kp9x\") on node \"crc\" DevicePath \"\"" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.516451 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.525065 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" (UID: "8ae16d21-1bb0-47c2-9c55-66bfb00ebd70"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.582679 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.617383 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.718686 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-utilities\") pod \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.718858 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-catalog-content\") pod \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.718931 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlkgt\" (UniqueName: \"kubernetes.io/projected/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-kube-api-access-zlkgt\") pod \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\" (UID: \"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb\") " Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.720039 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-utilities" (OuterVolumeSpecName: "utilities") pod "0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" (UID: "0cccc8c7-4f29-4d86-9463-e2c71d4aaafb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.724307 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-kube-api-access-zlkgt" (OuterVolumeSpecName: "kube-api-access-zlkgt") pod "0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" (UID: "0cccc8c7-4f29-4d86-9463-e2c71d4aaafb"). InnerVolumeSpecName "kube-api-access-zlkgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.763643 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" (UID: "0cccc8c7-4f29-4d86-9463-e2c71d4aaafb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.820732 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.820770 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:37:02 crc kubenswrapper[4944]: I1124 09:37:02.820781 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlkgt\" (UniqueName: \"kubernetes.io/projected/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb-kube-api-access-zlkgt\") on node \"crc\" DevicePath \"\"" Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.215203 4944 generic.go:334] "Generic (PLEG): container finished" podID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" containerID="c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790" exitCode=0 Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.215269 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b4pv4" Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.215285 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4pv4" event={"ID":"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb","Type":"ContainerDied","Data":"c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790"} Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.216272 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b4pv4" event={"ID":"0cccc8c7-4f29-4d86-9463-e2c71d4aaafb","Type":"ContainerDied","Data":"3a670fd63cf00f8634b101767e82dc650d852867b0206738918027c7ba66f19b"} Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.216318 4944 scope.go:117] "RemoveContainer" containerID="c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790" Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.216401 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w2n6" Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.238864 4944 scope.go:117] "RemoveContainer" containerID="9bf66e89bf8acb34bc7f9008a01b553149578db4b06877113e1b5da6db5c1b76" Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.248703 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5w2n6"] Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.253513 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5w2n6"] Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.262674 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b4pv4"] Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.270834 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b4pv4"] Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.277822 4944 scope.go:117] "RemoveContainer" containerID="93152006e2f6bd4f123ad5111eee2a52778e9d7b9c8388c76bea0064e888499e" Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.292544 4944 scope.go:117] "RemoveContainer" containerID="c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790" Nov 24 09:37:03 crc kubenswrapper[4944]: E1124 09:37:03.293073 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790\": container with ID starting with c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790 not found: ID does not exist" containerID="c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790" Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.293116 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790"} err="failed to get container status \"c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790\": rpc error: code = NotFound desc = could not find container \"c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790\": container with ID starting with c2997be9b9f746af0e349567f97ae284207c6ff7a66786503f10b4aa25679790 not found: ID does not exist" Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.293143 4944 scope.go:117] "RemoveContainer" containerID="9bf66e89bf8acb34bc7f9008a01b553149578db4b06877113e1b5da6db5c1b76" Nov 24 09:37:03 crc kubenswrapper[4944]: E1124 09:37:03.293512 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bf66e89bf8acb34bc7f9008a01b553149578db4b06877113e1b5da6db5c1b76\": container with ID starting with 9bf66e89bf8acb34bc7f9008a01b553149578db4b06877113e1b5da6db5c1b76 not found: ID does not exist" containerID="9bf66e89bf8acb34bc7f9008a01b553149578db4b06877113e1b5da6db5c1b76" Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.293551 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bf66e89bf8acb34bc7f9008a01b553149578db4b06877113e1b5da6db5c1b76"} err="failed to get container status \"9bf66e89bf8acb34bc7f9008a01b553149578db4b06877113e1b5da6db5c1b76\": rpc error: code = NotFound desc = could not find container \"9bf66e89bf8acb34bc7f9008a01b553149578db4b06877113e1b5da6db5c1b76\": container with ID starting with 9bf66e89bf8acb34bc7f9008a01b553149578db4b06877113e1b5da6db5c1b76 not found: ID does not exist" Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.293578 4944 scope.go:117] "RemoveContainer" containerID="93152006e2f6bd4f123ad5111eee2a52778e9d7b9c8388c76bea0064e888499e" Nov 24 09:37:03 crc kubenswrapper[4944]: E1124 09:37:03.293952 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93152006e2f6bd4f123ad5111eee2a52778e9d7b9c8388c76bea0064e888499e\": container with ID starting with 93152006e2f6bd4f123ad5111eee2a52778e9d7b9c8388c76bea0064e888499e not found: ID does not exist" containerID="93152006e2f6bd4f123ad5111eee2a52778e9d7b9c8388c76bea0064e888499e" Nov 24 09:37:03 crc kubenswrapper[4944]: I1124 09:37:03.293982 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93152006e2f6bd4f123ad5111eee2a52778e9d7b9c8388c76bea0064e888499e"} err="failed to get container status \"93152006e2f6bd4f123ad5111eee2a52778e9d7b9c8388c76bea0064e888499e\": rpc error: code = NotFound desc = could not find container \"93152006e2f6bd4f123ad5111eee2a52778e9d7b9c8388c76bea0064e888499e\": container with ID starting with 93152006e2f6bd4f123ad5111eee2a52778e9d7b9c8388c76bea0064e888499e not found: ID does not exist" Nov 24 09:37:04 crc kubenswrapper[4944]: I1124 09:37:04.286678 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" path="/var/lib/kubelet/pods/0cccc8c7-4f29-4d86-9463-e2c71d4aaafb/volumes" Nov 24 09:37:04 crc kubenswrapper[4944]: I1124 09:37:04.288193 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" path="/var/lib/kubelet/pods/8ae16d21-1bb0-47c2-9c55-66bfb00ebd70/volumes" Nov 24 09:37:23 crc kubenswrapper[4944]: I1124 09:37:23.549012 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:37:23 crc kubenswrapper[4944]: I1124 09:37:23.549600 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:37:53 crc kubenswrapper[4944]: I1124 09:37:53.548432 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:37:53 crc kubenswrapper[4944]: I1124 09:37:53.548982 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:37:53 crc kubenswrapper[4944]: I1124 09:37:53.549029 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:37:53 crc kubenswrapper[4944]: I1124 09:37:53.549620 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f7d69e51c92939bdcd5b3f695ca62927d0f1b11c6749722c6131860a7f19cb18"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:37:53 crc kubenswrapper[4944]: I1124 09:37:53.549662 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://f7d69e51c92939bdcd5b3f695ca62927d0f1b11c6749722c6131860a7f19cb18" gracePeriod=600 Nov 24 09:37:54 crc kubenswrapper[4944]: I1124 09:37:54.576372 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="f7d69e51c92939bdcd5b3f695ca62927d0f1b11c6749722c6131860a7f19cb18" exitCode=0 Nov 24 09:37:54 crc kubenswrapper[4944]: I1124 09:37:54.576449 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"f7d69e51c92939bdcd5b3f695ca62927d0f1b11c6749722c6131860a7f19cb18"} Nov 24 09:37:54 crc kubenswrapper[4944]: I1124 09:37:54.576733 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859"} Nov 24 09:37:54 crc kubenswrapper[4944]: I1124 09:37:54.576761 4944 scope.go:117] "RemoveContainer" containerID="0204116cb52f453bbaa78b7cfc5f12c8fb9af61d36f58c99d39064bae1b1e3d7" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.164242 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xz9hq"] Nov 24 09:38:37 crc kubenswrapper[4944]: E1124 09:38:37.166355 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" containerName="registry-server" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.166375 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" containerName="registry-server" Nov 24 09:38:37 crc kubenswrapper[4944]: E1124 09:38:37.166397 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" containerName="extract-utilities" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.166406 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" containerName="extract-utilities" Nov 24 09:38:37 crc kubenswrapper[4944]: E1124 09:38:37.166426 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" containerName="extract-content" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.166434 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" containerName="extract-content" Nov 24 09:38:37 crc kubenswrapper[4944]: E1124 09:38:37.166447 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" containerName="extract-utilities" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.166454 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" containerName="extract-utilities" Nov 24 09:38:37 crc kubenswrapper[4944]: E1124 09:38:37.166467 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" containerName="extract-content" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.166474 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" containerName="extract-content" Nov 24 09:38:37 crc kubenswrapper[4944]: E1124 09:38:37.166491 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" containerName="registry-server" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.166497 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" containerName="registry-server" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.166652 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cccc8c7-4f29-4d86-9463-e2c71d4aaafb" containerName="registry-server" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.166676 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae16d21-1bb0-47c2-9c55-66bfb00ebd70" containerName="registry-server" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.167855 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.191232 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xz9hq"] Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.208380 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztk8p\" (UniqueName: \"kubernetes.io/projected/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-kube-api-access-ztk8p\") pod \"redhat-marketplace-xz9hq\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.208550 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-utilities\") pod \"redhat-marketplace-xz9hq\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.208690 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-catalog-content\") pod \"redhat-marketplace-xz9hq\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.309737 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztk8p\" (UniqueName: \"kubernetes.io/projected/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-kube-api-access-ztk8p\") pod \"redhat-marketplace-xz9hq\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.309811 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-utilities\") pod \"redhat-marketplace-xz9hq\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.309856 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-catalog-content\") pod \"redhat-marketplace-xz9hq\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.310539 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-utilities\") pod \"redhat-marketplace-xz9hq\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.310586 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-catalog-content\") pod \"redhat-marketplace-xz9hq\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.331552 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztk8p\" (UniqueName: \"kubernetes.io/projected/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-kube-api-access-ztk8p\") pod \"redhat-marketplace-xz9hq\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.491067 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.762331 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xz9hq"] Nov 24 09:38:37 crc kubenswrapper[4944]: I1124 09:38:37.872757 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xz9hq" event={"ID":"412978e0-3a9e-42a4-8c5a-ec60a39b6af6","Type":"ContainerStarted","Data":"9e2d0538b47764147b2d8d89def7b4b86f681e88880ba78fc2567ca746bdf8af"} Nov 24 09:38:38 crc kubenswrapper[4944]: I1124 09:38:38.880770 4944 generic.go:334] "Generic (PLEG): container finished" podID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" containerID="2530378d312d547ed604ccc84d0336c88d9c17271f6c1a0f3a94876fadee0d22" exitCode=0 Nov 24 09:38:38 crc kubenswrapper[4944]: I1124 09:38:38.880812 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xz9hq" event={"ID":"412978e0-3a9e-42a4-8c5a-ec60a39b6af6","Type":"ContainerDied","Data":"2530378d312d547ed604ccc84d0336c88d9c17271f6c1a0f3a94876fadee0d22"} Nov 24 09:38:40 crc kubenswrapper[4944]: I1124 09:38:40.894734 4944 generic.go:334] "Generic (PLEG): container finished" podID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" containerID="4973fb741fe8cdbf364f9fe6989dd5d62dbe04af8a02bd17813d756b8223ec01" exitCode=0 Nov 24 09:38:40 crc kubenswrapper[4944]: I1124 09:38:40.894854 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xz9hq" event={"ID":"412978e0-3a9e-42a4-8c5a-ec60a39b6af6","Type":"ContainerDied","Data":"4973fb741fe8cdbf364f9fe6989dd5d62dbe04af8a02bd17813d756b8223ec01"} Nov 24 09:38:41 crc kubenswrapper[4944]: I1124 09:38:41.905505 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xz9hq" event={"ID":"412978e0-3a9e-42a4-8c5a-ec60a39b6af6","Type":"ContainerStarted","Data":"cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532"} Nov 24 09:38:41 crc kubenswrapper[4944]: I1124 09:38:41.928784 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xz9hq" podStartSLOduration=2.329890124 podStartE2EDuration="4.928764054s" podCreationTimestamp="2025-11-24 09:38:37 +0000 UTC" firstStartedPulling="2025-11-24 09:38:38.882757537 +0000 UTC m=+2779.417197999" lastFinishedPulling="2025-11-24 09:38:41.481631467 +0000 UTC m=+2782.016071929" observedRunningTime="2025-11-24 09:38:41.927263105 +0000 UTC m=+2782.461703567" watchObservedRunningTime="2025-11-24 09:38:41.928764054 +0000 UTC m=+2782.463204536" Nov 24 09:38:47 crc kubenswrapper[4944]: I1124 09:38:47.491271 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:47 crc kubenswrapper[4944]: I1124 09:38:47.492335 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:47 crc kubenswrapper[4944]: I1124 09:38:47.534928 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:47 crc kubenswrapper[4944]: I1124 09:38:47.991364 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:48 crc kubenswrapper[4944]: I1124 09:38:48.080715 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xz9hq"] Nov 24 09:38:49 crc kubenswrapper[4944]: I1124 09:38:49.959152 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xz9hq" podUID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" containerName="registry-server" containerID="cri-o://cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532" gracePeriod=2 Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.324915 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.376528 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztk8p\" (UniqueName: \"kubernetes.io/projected/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-kube-api-access-ztk8p\") pod \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.376573 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-utilities\") pod \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.376610 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-catalog-content\") pod \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\" (UID: \"412978e0-3a9e-42a4-8c5a-ec60a39b6af6\") " Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.377748 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-utilities" (OuterVolumeSpecName: "utilities") pod "412978e0-3a9e-42a4-8c5a-ec60a39b6af6" (UID: "412978e0-3a9e-42a4-8c5a-ec60a39b6af6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.383264 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-kube-api-access-ztk8p" (OuterVolumeSpecName: "kube-api-access-ztk8p") pod "412978e0-3a9e-42a4-8c5a-ec60a39b6af6" (UID: "412978e0-3a9e-42a4-8c5a-ec60a39b6af6"). InnerVolumeSpecName "kube-api-access-ztk8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.395543 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "412978e0-3a9e-42a4-8c5a-ec60a39b6af6" (UID: "412978e0-3a9e-42a4-8c5a-ec60a39b6af6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.478737 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.478768 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztk8p\" (UniqueName: \"kubernetes.io/projected/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-kube-api-access-ztk8p\") on node \"crc\" DevicePath \"\"" Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.478779 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412978e0-3a9e-42a4-8c5a-ec60a39b6af6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.967579 4944 generic.go:334] "Generic (PLEG): container finished" podID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" containerID="cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532" exitCode=0 Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.967628 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xz9hq" Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.967633 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xz9hq" event={"ID":"412978e0-3a9e-42a4-8c5a-ec60a39b6af6","Type":"ContainerDied","Data":"cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532"} Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.968080 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xz9hq" event={"ID":"412978e0-3a9e-42a4-8c5a-ec60a39b6af6","Type":"ContainerDied","Data":"9e2d0538b47764147b2d8d89def7b4b86f681e88880ba78fc2567ca746bdf8af"} Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.968111 4944 scope.go:117] "RemoveContainer" containerID="cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532" Nov 24 09:38:50 crc kubenswrapper[4944]: I1124 09:38:50.989591 4944 scope.go:117] "RemoveContainer" containerID="4973fb741fe8cdbf364f9fe6989dd5d62dbe04af8a02bd17813d756b8223ec01" Nov 24 09:38:51 crc kubenswrapper[4944]: I1124 09:38:51.011287 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xz9hq"] Nov 24 09:38:51 crc kubenswrapper[4944]: I1124 09:38:51.015801 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xz9hq"] Nov 24 09:38:51 crc kubenswrapper[4944]: I1124 09:38:51.043757 4944 scope.go:117] "RemoveContainer" containerID="2530378d312d547ed604ccc84d0336c88d9c17271f6c1a0f3a94876fadee0d22" Nov 24 09:38:51 crc kubenswrapper[4944]: I1124 09:38:51.059493 4944 scope.go:117] "RemoveContainer" containerID="cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532" Nov 24 09:38:51 crc kubenswrapper[4944]: E1124 09:38:51.060108 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532\": container with ID starting with cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532 not found: ID does not exist" containerID="cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532" Nov 24 09:38:51 crc kubenswrapper[4944]: I1124 09:38:51.060160 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532"} err="failed to get container status \"cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532\": rpc error: code = NotFound desc = could not find container \"cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532\": container with ID starting with cd2b1d75aa5889520d8d03ebe9dfe0df0f8dea1716d81c9bc2c6ff88cfff1532 not found: ID does not exist" Nov 24 09:38:51 crc kubenswrapper[4944]: I1124 09:38:51.060194 4944 scope.go:117] "RemoveContainer" containerID="4973fb741fe8cdbf364f9fe6989dd5d62dbe04af8a02bd17813d756b8223ec01" Nov 24 09:38:51 crc kubenswrapper[4944]: E1124 09:38:51.060650 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4973fb741fe8cdbf364f9fe6989dd5d62dbe04af8a02bd17813d756b8223ec01\": container with ID starting with 4973fb741fe8cdbf364f9fe6989dd5d62dbe04af8a02bd17813d756b8223ec01 not found: ID does not exist" containerID="4973fb741fe8cdbf364f9fe6989dd5d62dbe04af8a02bd17813d756b8223ec01" Nov 24 09:38:51 crc kubenswrapper[4944]: I1124 09:38:51.060675 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4973fb741fe8cdbf364f9fe6989dd5d62dbe04af8a02bd17813d756b8223ec01"} err="failed to get container status \"4973fb741fe8cdbf364f9fe6989dd5d62dbe04af8a02bd17813d756b8223ec01\": rpc error: code = NotFound desc = could not find container \"4973fb741fe8cdbf364f9fe6989dd5d62dbe04af8a02bd17813d756b8223ec01\": container with ID starting with 4973fb741fe8cdbf364f9fe6989dd5d62dbe04af8a02bd17813d756b8223ec01 not found: ID does not exist" Nov 24 09:38:51 crc kubenswrapper[4944]: I1124 09:38:51.060690 4944 scope.go:117] "RemoveContainer" containerID="2530378d312d547ed604ccc84d0336c88d9c17271f6c1a0f3a94876fadee0d22" Nov 24 09:38:51 crc kubenswrapper[4944]: E1124 09:38:51.060988 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2530378d312d547ed604ccc84d0336c88d9c17271f6c1a0f3a94876fadee0d22\": container with ID starting with 2530378d312d547ed604ccc84d0336c88d9c17271f6c1a0f3a94876fadee0d22 not found: ID does not exist" containerID="2530378d312d547ed604ccc84d0336c88d9c17271f6c1a0f3a94876fadee0d22" Nov 24 09:38:51 crc kubenswrapper[4944]: I1124 09:38:51.061028 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2530378d312d547ed604ccc84d0336c88d9c17271f6c1a0f3a94876fadee0d22"} err="failed to get container status \"2530378d312d547ed604ccc84d0336c88d9c17271f6c1a0f3a94876fadee0d22\": rpc error: code = NotFound desc = could not find container \"2530378d312d547ed604ccc84d0336c88d9c17271f6c1a0f3a94876fadee0d22\": container with ID starting with 2530378d312d547ed604ccc84d0336c88d9c17271f6c1a0f3a94876fadee0d22 not found: ID does not exist" Nov 24 09:38:52 crc kubenswrapper[4944]: I1124 09:38:52.285364 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" path="/var/lib/kubelet/pods/412978e0-3a9e-42a4-8c5a-ec60a39b6af6/volumes" Nov 24 09:39:53 crc kubenswrapper[4944]: I1124 09:39:53.549463 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:39:53 crc kubenswrapper[4944]: I1124 09:39:53.550724 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:40:23 crc kubenswrapper[4944]: I1124 09:40:23.548969 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:40:23 crc kubenswrapper[4944]: I1124 09:40:23.549761 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.352936 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9pz44"] Nov 24 09:40:49 crc kubenswrapper[4944]: E1124 09:40:49.353847 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" containerName="extract-utilities" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.353864 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" containerName="extract-utilities" Nov 24 09:40:49 crc kubenswrapper[4944]: E1124 09:40:49.353888 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" containerName="extract-content" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.353896 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" containerName="extract-content" Nov 24 09:40:49 crc kubenswrapper[4944]: E1124 09:40:49.353911 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" containerName="registry-server" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.353919 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" containerName="registry-server" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.354217 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="412978e0-3a9e-42a4-8c5a-ec60a39b6af6" containerName="registry-server" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.356979 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.366089 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9pz44"] Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.432380 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8rlg\" (UniqueName: \"kubernetes.io/projected/18a72a61-caea-40ec-9dc9-da4a5cab1a24-kube-api-access-q8rlg\") pod \"community-operators-9pz44\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.432433 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-utilities\") pod \"community-operators-9pz44\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.432470 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-catalog-content\") pod \"community-operators-9pz44\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.533770 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8rlg\" (UniqueName: \"kubernetes.io/projected/18a72a61-caea-40ec-9dc9-da4a5cab1a24-kube-api-access-q8rlg\") pod \"community-operators-9pz44\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.534160 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-utilities\") pod \"community-operators-9pz44\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.534212 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-catalog-content\") pod \"community-operators-9pz44\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.534696 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-utilities\") pod \"community-operators-9pz44\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.534770 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-catalog-content\") pod \"community-operators-9pz44\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.559222 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8rlg\" (UniqueName: \"kubernetes.io/projected/18a72a61-caea-40ec-9dc9-da4a5cab1a24-kube-api-access-q8rlg\") pod \"community-operators-9pz44\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:49 crc kubenswrapper[4944]: I1124 09:40:49.683883 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:50 crc kubenswrapper[4944]: I1124 09:40:50.164727 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9pz44"] Nov 24 09:40:50 crc kubenswrapper[4944]: I1124 09:40:50.766580 4944 generic.go:334] "Generic (PLEG): container finished" podID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" containerID="dbfefce90c94699ad5972780e79ae4b8007ed0c23a661a91e772fb412c59de2a" exitCode=0 Nov 24 09:40:50 crc kubenswrapper[4944]: I1124 09:40:50.766620 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pz44" event={"ID":"18a72a61-caea-40ec-9dc9-da4a5cab1a24","Type":"ContainerDied","Data":"dbfefce90c94699ad5972780e79ae4b8007ed0c23a661a91e772fb412c59de2a"} Nov 24 09:40:50 crc kubenswrapper[4944]: I1124 09:40:50.766647 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pz44" event={"ID":"18a72a61-caea-40ec-9dc9-da4a5cab1a24","Type":"ContainerStarted","Data":"fb62940d4c9fb1b287a73f56d22fb8b6e860c00d1735a05e7fa04512fd9b0f99"} Nov 24 09:40:51 crc kubenswrapper[4944]: I1124 09:40:51.779706 4944 generic.go:334] "Generic (PLEG): container finished" podID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" containerID="0688cdf3b1091a6b38f21ba269de0f69ae27a98308a56a8798be523bbf2b89f3" exitCode=0 Nov 24 09:40:51 crc kubenswrapper[4944]: I1124 09:40:51.779891 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pz44" event={"ID":"18a72a61-caea-40ec-9dc9-da4a5cab1a24","Type":"ContainerDied","Data":"0688cdf3b1091a6b38f21ba269de0f69ae27a98308a56a8798be523bbf2b89f3"} Nov 24 09:40:52 crc kubenswrapper[4944]: I1124 09:40:52.805788 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pz44" event={"ID":"18a72a61-caea-40ec-9dc9-da4a5cab1a24","Type":"ContainerStarted","Data":"5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74"} Nov 24 09:40:52 crc kubenswrapper[4944]: I1124 09:40:52.826794 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9pz44" podStartSLOduration=2.344707356 podStartE2EDuration="3.826770199s" podCreationTimestamp="2025-11-24 09:40:49 +0000 UTC" firstStartedPulling="2025-11-24 09:40:50.768163355 +0000 UTC m=+2911.302603817" lastFinishedPulling="2025-11-24 09:40:52.250226198 +0000 UTC m=+2912.784666660" observedRunningTime="2025-11-24 09:40:52.823165855 +0000 UTC m=+2913.357606317" watchObservedRunningTime="2025-11-24 09:40:52.826770199 +0000 UTC m=+2913.361210651" Nov 24 09:40:53 crc kubenswrapper[4944]: I1124 09:40:53.548940 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:40:53 crc kubenswrapper[4944]: I1124 09:40:53.549011 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:40:53 crc kubenswrapper[4944]: I1124 09:40:53.549077 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:40:53 crc kubenswrapper[4944]: I1124 09:40:53.549655 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:40:53 crc kubenswrapper[4944]: I1124 09:40:53.549711 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" gracePeriod=600 Nov 24 09:40:53 crc kubenswrapper[4944]: E1124 09:40:53.678476 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:40:53 crc kubenswrapper[4944]: I1124 09:40:53.813908 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" exitCode=0 Nov 24 09:40:53 crc kubenswrapper[4944]: I1124 09:40:53.813988 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859"} Nov 24 09:40:53 crc kubenswrapper[4944]: I1124 09:40:53.814064 4944 scope.go:117] "RemoveContainer" containerID="f7d69e51c92939bdcd5b3f695ca62927d0f1b11c6749722c6131860a7f19cb18" Nov 24 09:40:53 crc kubenswrapper[4944]: I1124 09:40:53.814626 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:40:53 crc kubenswrapper[4944]: E1124 09:40:53.814892 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:40:59 crc kubenswrapper[4944]: I1124 09:40:59.684321 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:59 crc kubenswrapper[4944]: I1124 09:40:59.684886 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:59 crc kubenswrapper[4944]: I1124 09:40:59.728783 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:59 crc kubenswrapper[4944]: I1124 09:40:59.901238 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:40:59 crc kubenswrapper[4944]: I1124 09:40:59.957639 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9pz44"] Nov 24 09:41:01 crc kubenswrapper[4944]: I1124 09:41:01.874965 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9pz44" podUID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" containerName="registry-server" containerID="cri-o://5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74" gracePeriod=2 Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.797074 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.883694 4944 generic.go:334] "Generic (PLEG): container finished" podID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" containerID="5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74" exitCode=0 Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.883755 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pz44" event={"ID":"18a72a61-caea-40ec-9dc9-da4a5cab1a24","Type":"ContainerDied","Data":"5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74"} Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.883806 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9pz44" event={"ID":"18a72a61-caea-40ec-9dc9-da4a5cab1a24","Type":"ContainerDied","Data":"fb62940d4c9fb1b287a73f56d22fb8b6e860c00d1735a05e7fa04512fd9b0f99"} Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.883825 4944 scope.go:117] "RemoveContainer" containerID="5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.883837 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9pz44" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.899170 4944 scope.go:117] "RemoveContainer" containerID="0688cdf3b1091a6b38f21ba269de0f69ae27a98308a56a8798be523bbf2b89f3" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.915790 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-catalog-content\") pod \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.915869 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8rlg\" (UniqueName: \"kubernetes.io/projected/18a72a61-caea-40ec-9dc9-da4a5cab1a24-kube-api-access-q8rlg\") pod \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.915927 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-utilities\") pod \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\" (UID: \"18a72a61-caea-40ec-9dc9-da4a5cab1a24\") " Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.916693 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-utilities" (OuterVolumeSpecName: "utilities") pod "18a72a61-caea-40ec-9dc9-da4a5cab1a24" (UID: "18a72a61-caea-40ec-9dc9-da4a5cab1a24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.918746 4944 scope.go:117] "RemoveContainer" containerID="dbfefce90c94699ad5972780e79ae4b8007ed0c23a661a91e772fb412c59de2a" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.922376 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a72a61-caea-40ec-9dc9-da4a5cab1a24-kube-api-access-q8rlg" (OuterVolumeSpecName: "kube-api-access-q8rlg") pod "18a72a61-caea-40ec-9dc9-da4a5cab1a24" (UID: "18a72a61-caea-40ec-9dc9-da4a5cab1a24"). InnerVolumeSpecName "kube-api-access-q8rlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.964536 4944 scope.go:117] "RemoveContainer" containerID="5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74" Nov 24 09:41:02 crc kubenswrapper[4944]: E1124 09:41:02.965090 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74\": container with ID starting with 5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74 not found: ID does not exist" containerID="5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.965123 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74"} err="failed to get container status \"5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74\": rpc error: code = NotFound desc = could not find container \"5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74\": container with ID starting with 5f59a2ce7acda63d76223bee435d955783109077b42943f9fbfefc7b13906f74 not found: ID does not exist" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.965165 4944 scope.go:117] "RemoveContainer" containerID="0688cdf3b1091a6b38f21ba269de0f69ae27a98308a56a8798be523bbf2b89f3" Nov 24 09:41:02 crc kubenswrapper[4944]: E1124 09:41:02.965731 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0688cdf3b1091a6b38f21ba269de0f69ae27a98308a56a8798be523bbf2b89f3\": container with ID starting with 0688cdf3b1091a6b38f21ba269de0f69ae27a98308a56a8798be523bbf2b89f3 not found: ID does not exist" containerID="0688cdf3b1091a6b38f21ba269de0f69ae27a98308a56a8798be523bbf2b89f3" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.965762 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0688cdf3b1091a6b38f21ba269de0f69ae27a98308a56a8798be523bbf2b89f3"} err="failed to get container status \"0688cdf3b1091a6b38f21ba269de0f69ae27a98308a56a8798be523bbf2b89f3\": rpc error: code = NotFound desc = could not find container \"0688cdf3b1091a6b38f21ba269de0f69ae27a98308a56a8798be523bbf2b89f3\": container with ID starting with 0688cdf3b1091a6b38f21ba269de0f69ae27a98308a56a8798be523bbf2b89f3 not found: ID does not exist" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.965780 4944 scope.go:117] "RemoveContainer" containerID="dbfefce90c94699ad5972780e79ae4b8007ed0c23a661a91e772fb412c59de2a" Nov 24 09:41:02 crc kubenswrapper[4944]: E1124 09:41:02.966156 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbfefce90c94699ad5972780e79ae4b8007ed0c23a661a91e772fb412c59de2a\": container with ID starting with dbfefce90c94699ad5972780e79ae4b8007ed0c23a661a91e772fb412c59de2a not found: ID does not exist" containerID="dbfefce90c94699ad5972780e79ae4b8007ed0c23a661a91e772fb412c59de2a" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.966201 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbfefce90c94699ad5972780e79ae4b8007ed0c23a661a91e772fb412c59de2a"} err="failed to get container status \"dbfefce90c94699ad5972780e79ae4b8007ed0c23a661a91e772fb412c59de2a\": rpc error: code = NotFound desc = could not find container \"dbfefce90c94699ad5972780e79ae4b8007ed0c23a661a91e772fb412c59de2a\": container with ID starting with dbfefce90c94699ad5972780e79ae4b8007ed0c23a661a91e772fb412c59de2a not found: ID does not exist" Nov 24 09:41:02 crc kubenswrapper[4944]: I1124 09:41:02.972317 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18a72a61-caea-40ec-9dc9-da4a5cab1a24" (UID: "18a72a61-caea-40ec-9dc9-da4a5cab1a24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:41:03 crc kubenswrapper[4944]: I1124 09:41:03.017101 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8rlg\" (UniqueName: \"kubernetes.io/projected/18a72a61-caea-40ec-9dc9-da4a5cab1a24-kube-api-access-q8rlg\") on node \"crc\" DevicePath \"\"" Nov 24 09:41:03 crc kubenswrapper[4944]: I1124 09:41:03.017137 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:41:03 crc kubenswrapper[4944]: I1124 09:41:03.017149 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a72a61-caea-40ec-9dc9-da4a5cab1a24-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:41:03 crc kubenswrapper[4944]: I1124 09:41:03.219609 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9pz44"] Nov 24 09:41:03 crc kubenswrapper[4944]: I1124 09:41:03.231724 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9pz44"] Nov 24 09:41:04 crc kubenswrapper[4944]: I1124 09:41:04.293746 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" path="/var/lib/kubelet/pods/18a72a61-caea-40ec-9dc9-da4a5cab1a24/volumes" Nov 24 09:41:09 crc kubenswrapper[4944]: I1124 09:41:09.276935 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:41:09 crc kubenswrapper[4944]: E1124 09:41:09.278252 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:41:23 crc kubenswrapper[4944]: I1124 09:41:23.278324 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:41:23 crc kubenswrapper[4944]: E1124 09:41:23.279017 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:41:36 crc kubenswrapper[4944]: I1124 09:41:36.286390 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:41:36 crc kubenswrapper[4944]: E1124 09:41:36.287410 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:41:48 crc kubenswrapper[4944]: I1124 09:41:48.277531 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:41:48 crc kubenswrapper[4944]: E1124 09:41:48.278411 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:42:00 crc kubenswrapper[4944]: I1124 09:42:00.283618 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:42:00 crc kubenswrapper[4944]: E1124 09:42:00.284387 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:42:11 crc kubenswrapper[4944]: I1124 09:42:11.277368 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:42:11 crc kubenswrapper[4944]: E1124 09:42:11.278170 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:42:26 crc kubenswrapper[4944]: I1124 09:42:26.276794 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:42:26 crc kubenswrapper[4944]: E1124 09:42:26.277585 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:42:41 crc kubenswrapper[4944]: I1124 09:42:41.276640 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:42:41 crc kubenswrapper[4944]: E1124 09:42:41.277351 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:42:54 crc kubenswrapper[4944]: I1124 09:42:54.277146 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:42:54 crc kubenswrapper[4944]: E1124 09:42:54.278238 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:43:08 crc kubenswrapper[4944]: I1124 09:43:08.277263 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:43:08 crc kubenswrapper[4944]: E1124 09:43:08.278202 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:43:23 crc kubenswrapper[4944]: I1124 09:43:23.277341 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:43:23 crc kubenswrapper[4944]: E1124 09:43:23.278253 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:43:29 crc kubenswrapper[4944]: I1124 09:43:29.232614 4944 scope.go:117] "RemoveContainer" containerID="7b0cbfe239499374f05bf13664afb0db758d92bb13bcb9d38f5d01da5287e4c6" Nov 24 09:43:29 crc kubenswrapper[4944]: I1124 09:43:29.280252 4944 scope.go:117] "RemoveContainer" containerID="8c83d7c2d3ba033f24b3fbde739bb2b3705727d3ce88e64acfcf7083bbff3bf0" Nov 24 09:43:29 crc kubenswrapper[4944]: I1124 09:43:29.301723 4944 scope.go:117] "RemoveContainer" containerID="071afdaeafc8c930018a27809e8a431ecd7ace12f0e6d931b63fc811d8fe5c5e" Nov 24 09:43:34 crc kubenswrapper[4944]: I1124 09:43:34.277343 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:43:34 crc kubenswrapper[4944]: E1124 09:43:34.277821 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:43:48 crc kubenswrapper[4944]: I1124 09:43:48.276585 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:43:48 crc kubenswrapper[4944]: E1124 09:43:48.277250 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:44:02 crc kubenswrapper[4944]: I1124 09:44:02.277282 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:44:02 crc kubenswrapper[4944]: E1124 09:44:02.278160 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:44:13 crc kubenswrapper[4944]: I1124 09:44:13.276551 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:44:13 crc kubenswrapper[4944]: E1124 09:44:13.277202 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:44:24 crc kubenswrapper[4944]: I1124 09:44:24.276737 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:44:24 crc kubenswrapper[4944]: E1124 09:44:24.277602 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:44:35 crc kubenswrapper[4944]: I1124 09:44:35.276705 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:44:35 crc kubenswrapper[4944]: E1124 09:44:35.277485 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:44:48 crc kubenswrapper[4944]: I1124 09:44:48.277086 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:44:48 crc kubenswrapper[4944]: E1124 09:44:48.277790 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.188710 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c"] Nov 24 09:45:00 crc kubenswrapper[4944]: E1124 09:45:00.189560 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" containerName="extract-content" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.189579 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" containerName="extract-content" Nov 24 09:45:00 crc kubenswrapper[4944]: E1124 09:45:00.189620 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" containerName="extract-utilities" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.189628 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" containerName="extract-utilities" Nov 24 09:45:00 crc kubenswrapper[4944]: E1124 09:45:00.189642 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" containerName="registry-server" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.189649 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" containerName="registry-server" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.189832 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a72a61-caea-40ec-9dc9-da4a5cab1a24" containerName="registry-server" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.190407 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.195260 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.195521 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.209027 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c"] Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.308954 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ng9w\" (UniqueName: \"kubernetes.io/projected/3540edcd-cfcd-41eb-b351-9e749e87ce0b-kube-api-access-5ng9w\") pod \"collect-profiles-29399625-9tg4c\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.309069 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3540edcd-cfcd-41eb-b351-9e749e87ce0b-secret-volume\") pod \"collect-profiles-29399625-9tg4c\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.309097 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3540edcd-cfcd-41eb-b351-9e749e87ce0b-config-volume\") pod \"collect-profiles-29399625-9tg4c\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.410468 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3540edcd-cfcd-41eb-b351-9e749e87ce0b-secret-volume\") pod \"collect-profiles-29399625-9tg4c\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.410515 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3540edcd-cfcd-41eb-b351-9e749e87ce0b-config-volume\") pod \"collect-profiles-29399625-9tg4c\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.410546 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ng9w\" (UniqueName: \"kubernetes.io/projected/3540edcd-cfcd-41eb-b351-9e749e87ce0b-kube-api-access-5ng9w\") pod \"collect-profiles-29399625-9tg4c\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.411799 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3540edcd-cfcd-41eb-b351-9e749e87ce0b-config-volume\") pod \"collect-profiles-29399625-9tg4c\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.416190 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3540edcd-cfcd-41eb-b351-9e749e87ce0b-secret-volume\") pod \"collect-profiles-29399625-9tg4c\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.426173 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ng9w\" (UniqueName: \"kubernetes.io/projected/3540edcd-cfcd-41eb-b351-9e749e87ce0b-kube-api-access-5ng9w\") pod \"collect-profiles-29399625-9tg4c\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.511596 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:00 crc kubenswrapper[4944]: I1124 09:45:00.918746 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c"] Nov 24 09:45:00 crc kubenswrapper[4944]: W1124 09:45:00.930226 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3540edcd_cfcd_41eb_b351_9e749e87ce0b.slice/crio-80aaf9800a39cb18eb7a202b1b65737b01b5bf5607a181cf027fb10e5121d095 WatchSource:0}: Error finding container 80aaf9800a39cb18eb7a202b1b65737b01b5bf5607a181cf027fb10e5121d095: Status 404 returned error can't find the container with id 80aaf9800a39cb18eb7a202b1b65737b01b5bf5607a181cf027fb10e5121d095 Nov 24 09:45:01 crc kubenswrapper[4944]: I1124 09:45:01.768747 4944 generic.go:334] "Generic (PLEG): container finished" podID="3540edcd-cfcd-41eb-b351-9e749e87ce0b" containerID="ff0dbb3d0734e85266fb5db92d0779680ea8e630abbc2b0a439bc4195aaa4403" exitCode=0 Nov 24 09:45:01 crc kubenswrapper[4944]: I1124 09:45:01.768855 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" event={"ID":"3540edcd-cfcd-41eb-b351-9e749e87ce0b","Type":"ContainerDied","Data":"ff0dbb3d0734e85266fb5db92d0779680ea8e630abbc2b0a439bc4195aaa4403"} Nov 24 09:45:01 crc kubenswrapper[4944]: I1124 09:45:01.769169 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" event={"ID":"3540edcd-cfcd-41eb-b351-9e749e87ce0b","Type":"ContainerStarted","Data":"80aaf9800a39cb18eb7a202b1b65737b01b5bf5607a181cf027fb10e5121d095"} Nov 24 09:45:02 crc kubenswrapper[4944]: I1124 09:45:02.277112 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:45:02 crc kubenswrapper[4944]: E1124 09:45:02.277597 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.125895 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.251909 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3540edcd-cfcd-41eb-b351-9e749e87ce0b-secret-volume\") pod \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.251970 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ng9w\" (UniqueName: \"kubernetes.io/projected/3540edcd-cfcd-41eb-b351-9e749e87ce0b-kube-api-access-5ng9w\") pod \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.252010 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3540edcd-cfcd-41eb-b351-9e749e87ce0b-config-volume\") pod \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\" (UID: \"3540edcd-cfcd-41eb-b351-9e749e87ce0b\") " Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.252798 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3540edcd-cfcd-41eb-b351-9e749e87ce0b-config-volume" (OuterVolumeSpecName: "config-volume") pod "3540edcd-cfcd-41eb-b351-9e749e87ce0b" (UID: "3540edcd-cfcd-41eb-b351-9e749e87ce0b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.257333 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3540edcd-cfcd-41eb-b351-9e749e87ce0b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3540edcd-cfcd-41eb-b351-9e749e87ce0b" (UID: "3540edcd-cfcd-41eb-b351-9e749e87ce0b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.258206 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3540edcd-cfcd-41eb-b351-9e749e87ce0b-kube-api-access-5ng9w" (OuterVolumeSpecName: "kube-api-access-5ng9w") pod "3540edcd-cfcd-41eb-b351-9e749e87ce0b" (UID: "3540edcd-cfcd-41eb-b351-9e749e87ce0b"). InnerVolumeSpecName "kube-api-access-5ng9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.357902 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3540edcd-cfcd-41eb-b351-9e749e87ce0b-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.357934 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ng9w\" (UniqueName: \"kubernetes.io/projected/3540edcd-cfcd-41eb-b351-9e749e87ce0b-kube-api-access-5ng9w\") on node \"crc\" DevicePath \"\"" Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.357943 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3540edcd-cfcd-41eb-b351-9e749e87ce0b-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.786647 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" event={"ID":"3540edcd-cfcd-41eb-b351-9e749e87ce0b","Type":"ContainerDied","Data":"80aaf9800a39cb18eb7a202b1b65737b01b5bf5607a181cf027fb10e5121d095"} Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.786753 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c" Nov 24 09:45:03 crc kubenswrapper[4944]: I1124 09:45:03.787574 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80aaf9800a39cb18eb7a202b1b65737b01b5bf5607a181cf027fb10e5121d095" Nov 24 09:45:04 crc kubenswrapper[4944]: I1124 09:45:04.216096 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd"] Nov 24 09:45:04 crc kubenswrapper[4944]: I1124 09:45:04.221861 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399580-bwtnd"] Nov 24 09:45:04 crc kubenswrapper[4944]: I1124 09:45:04.288745 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac60169b-7886-4307-abca-27316f3c4d0f" path="/var/lib/kubelet/pods/ac60169b-7886-4307-abca-27316f3c4d0f/volumes" Nov 24 09:45:17 crc kubenswrapper[4944]: I1124 09:45:17.276993 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:45:17 crc kubenswrapper[4944]: E1124 09:45:17.277939 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:45:29 crc kubenswrapper[4944]: I1124 09:45:29.363559 4944 scope.go:117] "RemoveContainer" containerID="01a8d99f749f4ef70f2e5f5c87c64e59aa075551a45050e3ef9750c374dd16b1" Nov 24 09:45:31 crc kubenswrapper[4944]: I1124 09:45:31.276735 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:45:31 crc kubenswrapper[4944]: E1124 09:45:31.277788 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:45:44 crc kubenswrapper[4944]: I1124 09:45:44.277675 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:45:44 crc kubenswrapper[4944]: E1124 09:45:44.278923 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:45:59 crc kubenswrapper[4944]: I1124 09:45:59.276851 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:46:00 crc kubenswrapper[4944]: I1124 09:46:00.246960 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"de0b9680576b7802499acfa8e09920ec83923c1502e8d607ed1ab1120765d7ce"} Nov 24 09:48:23 crc kubenswrapper[4944]: I1124 09:48:23.548207 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:48:23 crc kubenswrapper[4944]: I1124 09:48:23.548820 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:48:26 crc kubenswrapper[4944]: I1124 09:48:26.895519 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t882v"] Nov 24 09:48:26 crc kubenswrapper[4944]: E1124 09:48:26.896668 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3540edcd-cfcd-41eb-b351-9e749e87ce0b" containerName="collect-profiles" Nov 24 09:48:26 crc kubenswrapper[4944]: I1124 09:48:26.896690 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3540edcd-cfcd-41eb-b351-9e749e87ce0b" containerName="collect-profiles" Nov 24 09:48:26 crc kubenswrapper[4944]: I1124 09:48:26.896896 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3540edcd-cfcd-41eb-b351-9e749e87ce0b" containerName="collect-profiles" Nov 24 09:48:26 crc kubenswrapper[4944]: I1124 09:48:26.898918 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:26 crc kubenswrapper[4944]: I1124 09:48:26.907956 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t882v"] Nov 24 09:48:26 crc kubenswrapper[4944]: I1124 09:48:26.948919 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-catalog-content\") pod \"certified-operators-t882v\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:26 crc kubenswrapper[4944]: I1124 09:48:26.948972 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb979\" (UniqueName: \"kubernetes.io/projected/d6a42739-d3f6-4a4f-a750-899b68adb4de-kube-api-access-jb979\") pod \"certified-operators-t882v\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:26 crc kubenswrapper[4944]: I1124 09:48:26.949026 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-utilities\") pod \"certified-operators-t882v\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:27 crc kubenswrapper[4944]: I1124 09:48:27.049803 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-catalog-content\") pod \"certified-operators-t882v\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:27 crc kubenswrapper[4944]: I1124 09:48:27.049880 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb979\" (UniqueName: \"kubernetes.io/projected/d6a42739-d3f6-4a4f-a750-899b68adb4de-kube-api-access-jb979\") pod \"certified-operators-t882v\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:27 crc kubenswrapper[4944]: I1124 09:48:27.049924 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-utilities\") pod \"certified-operators-t882v\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:27 crc kubenswrapper[4944]: I1124 09:48:27.050399 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-catalog-content\") pod \"certified-operators-t882v\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:27 crc kubenswrapper[4944]: I1124 09:48:27.050486 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-utilities\") pod \"certified-operators-t882v\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:27 crc kubenswrapper[4944]: I1124 09:48:27.076247 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb979\" (UniqueName: \"kubernetes.io/projected/d6a42739-d3f6-4a4f-a750-899b68adb4de-kube-api-access-jb979\") pod \"certified-operators-t882v\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:27 crc kubenswrapper[4944]: I1124 09:48:27.223412 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:27 crc kubenswrapper[4944]: I1124 09:48:27.684033 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t882v"] Nov 24 09:48:28 crc kubenswrapper[4944]: I1124 09:48:28.302510 4944 generic.go:334] "Generic (PLEG): container finished" podID="d6a42739-d3f6-4a4f-a750-899b68adb4de" containerID="9321673dbdb7d203b8f727fd8d077c524b7e30bfe87044a103c4fb6798b6f6f7" exitCode=0 Nov 24 09:48:28 crc kubenswrapper[4944]: I1124 09:48:28.302606 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t882v" event={"ID":"d6a42739-d3f6-4a4f-a750-899b68adb4de","Type":"ContainerDied","Data":"9321673dbdb7d203b8f727fd8d077c524b7e30bfe87044a103c4fb6798b6f6f7"} Nov 24 09:48:28 crc kubenswrapper[4944]: I1124 09:48:28.302811 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t882v" event={"ID":"d6a42739-d3f6-4a4f-a750-899b68adb4de","Type":"ContainerStarted","Data":"aacd727f30e4508a4e81c9480a51a6e679ff08f325bed48f161048384e179830"} Nov 24 09:48:28 crc kubenswrapper[4944]: I1124 09:48:28.305152 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.289222 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wf2rp"] Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.291354 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.306577 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wf2rp"] Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.331883 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t882v" event={"ID":"d6a42739-d3f6-4a4f-a750-899b68adb4de","Type":"ContainerStarted","Data":"eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1"} Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.397597 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwks7\" (UniqueName: \"kubernetes.io/projected/e92bfd19-27bd-413c-a718-c1279aff7dff-kube-api-access-vwks7\") pod \"redhat-operators-wf2rp\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.397650 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-catalog-content\") pod \"redhat-operators-wf2rp\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.397708 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-utilities\") pod \"redhat-operators-wf2rp\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.499448 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwks7\" (UniqueName: \"kubernetes.io/projected/e92bfd19-27bd-413c-a718-c1279aff7dff-kube-api-access-vwks7\") pod \"redhat-operators-wf2rp\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.499496 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-catalog-content\") pod \"redhat-operators-wf2rp\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.499524 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-utilities\") pod \"redhat-operators-wf2rp\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.499943 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-utilities\") pod \"redhat-operators-wf2rp\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.500014 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-catalog-content\") pod \"redhat-operators-wf2rp\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.517625 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwks7\" (UniqueName: \"kubernetes.io/projected/e92bfd19-27bd-413c-a718-c1279aff7dff-kube-api-access-vwks7\") pod \"redhat-operators-wf2rp\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:29 crc kubenswrapper[4944]: I1124 09:48:29.611090 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:30 crc kubenswrapper[4944]: I1124 09:48:30.097234 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wf2rp"] Nov 24 09:48:30 crc kubenswrapper[4944]: I1124 09:48:30.338669 4944 generic.go:334] "Generic (PLEG): container finished" podID="e92bfd19-27bd-413c-a718-c1279aff7dff" containerID="39acb028e22fa3e871226569a8ac103505af8919205b80646de63c9904346a66" exitCode=0 Nov 24 09:48:30 crc kubenswrapper[4944]: I1124 09:48:30.338885 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wf2rp" event={"ID":"e92bfd19-27bd-413c-a718-c1279aff7dff","Type":"ContainerDied","Data":"39acb028e22fa3e871226569a8ac103505af8919205b80646de63c9904346a66"} Nov 24 09:48:30 crc kubenswrapper[4944]: I1124 09:48:30.339076 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wf2rp" event={"ID":"e92bfd19-27bd-413c-a718-c1279aff7dff","Type":"ContainerStarted","Data":"a1fee63dc1bab3654d542f8b38fe702d9f45590269d7d165b16e0c4fac2e6de2"} Nov 24 09:48:30 crc kubenswrapper[4944]: I1124 09:48:30.341711 4944 generic.go:334] "Generic (PLEG): container finished" podID="d6a42739-d3f6-4a4f-a750-899b68adb4de" containerID="eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1" exitCode=0 Nov 24 09:48:30 crc kubenswrapper[4944]: I1124 09:48:30.341736 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t882v" event={"ID":"d6a42739-d3f6-4a4f-a750-899b68adb4de","Type":"ContainerDied","Data":"eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1"} Nov 24 09:48:31 crc kubenswrapper[4944]: I1124 09:48:31.348479 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wf2rp" event={"ID":"e92bfd19-27bd-413c-a718-c1279aff7dff","Type":"ContainerStarted","Data":"b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf"} Nov 24 09:48:31 crc kubenswrapper[4944]: I1124 09:48:31.351329 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t882v" event={"ID":"d6a42739-d3f6-4a4f-a750-899b68adb4de","Type":"ContainerStarted","Data":"27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717"} Nov 24 09:48:31 crc kubenswrapper[4944]: I1124 09:48:31.390368 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t882v" podStartSLOduration=2.956158796 podStartE2EDuration="5.390345491s" podCreationTimestamp="2025-11-24 09:48:26 +0000 UTC" firstStartedPulling="2025-11-24 09:48:28.304642739 +0000 UTC m=+3368.839083201" lastFinishedPulling="2025-11-24 09:48:30.738829434 +0000 UTC m=+3371.273269896" observedRunningTime="2025-11-24 09:48:31.389574196 +0000 UTC m=+3371.924014668" watchObservedRunningTime="2025-11-24 09:48:31.390345491 +0000 UTC m=+3371.924785953" Nov 24 09:48:32 crc kubenswrapper[4944]: I1124 09:48:32.361898 4944 generic.go:334] "Generic (PLEG): container finished" podID="e92bfd19-27bd-413c-a718-c1279aff7dff" containerID="b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf" exitCode=0 Nov 24 09:48:32 crc kubenswrapper[4944]: I1124 09:48:32.361963 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wf2rp" event={"ID":"e92bfd19-27bd-413c-a718-c1279aff7dff","Type":"ContainerDied","Data":"b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf"} Nov 24 09:48:33 crc kubenswrapper[4944]: I1124 09:48:33.371951 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wf2rp" event={"ID":"e92bfd19-27bd-413c-a718-c1279aff7dff","Type":"ContainerStarted","Data":"a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9"} Nov 24 09:48:33 crc kubenswrapper[4944]: I1124 09:48:33.390916 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wf2rp" podStartSLOduration=1.892912268 podStartE2EDuration="4.390899884s" podCreationTimestamp="2025-11-24 09:48:29 +0000 UTC" firstStartedPulling="2025-11-24 09:48:30.341095324 +0000 UTC m=+3370.875535786" lastFinishedPulling="2025-11-24 09:48:32.83908295 +0000 UTC m=+3373.373523402" observedRunningTime="2025-11-24 09:48:33.386797164 +0000 UTC m=+3373.921237626" watchObservedRunningTime="2025-11-24 09:48:33.390899884 +0000 UTC m=+3373.925340346" Nov 24 09:48:37 crc kubenswrapper[4944]: I1124 09:48:37.224769 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:37 crc kubenswrapper[4944]: I1124 09:48:37.225220 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:37 crc kubenswrapper[4944]: I1124 09:48:37.274643 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:37 crc kubenswrapper[4944]: I1124 09:48:37.438157 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:38 crc kubenswrapper[4944]: I1124 09:48:38.483170 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t882v"] Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.409287 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t882v" podUID="d6a42739-d3f6-4a4f-a750-899b68adb4de" containerName="registry-server" containerID="cri-o://27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717" gracePeriod=2 Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.611623 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.612825 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.657773 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.750003 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.886255 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jb979\" (UniqueName: \"kubernetes.io/projected/d6a42739-d3f6-4a4f-a750-899b68adb4de-kube-api-access-jb979\") pod \"d6a42739-d3f6-4a4f-a750-899b68adb4de\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.886356 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-utilities\") pod \"d6a42739-d3f6-4a4f-a750-899b68adb4de\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.886418 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-catalog-content\") pod \"d6a42739-d3f6-4a4f-a750-899b68adb4de\" (UID: \"d6a42739-d3f6-4a4f-a750-899b68adb4de\") " Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.887483 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-utilities" (OuterVolumeSpecName: "utilities") pod "d6a42739-d3f6-4a4f-a750-899b68adb4de" (UID: "d6a42739-d3f6-4a4f-a750-899b68adb4de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.891299 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a42739-d3f6-4a4f-a750-899b68adb4de-kube-api-access-jb979" (OuterVolumeSpecName: "kube-api-access-jb979") pod "d6a42739-d3f6-4a4f-a750-899b68adb4de" (UID: "d6a42739-d3f6-4a4f-a750-899b68adb4de"). InnerVolumeSpecName "kube-api-access-jb979". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.939021 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6a42739-d3f6-4a4f-a750-899b68adb4de" (UID: "d6a42739-d3f6-4a4f-a750-899b68adb4de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.989376 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.989432 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6a42739-d3f6-4a4f-a750-899b68adb4de-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:48:39 crc kubenswrapper[4944]: I1124 09:48:39.989449 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jb979\" (UniqueName: \"kubernetes.io/projected/d6a42739-d3f6-4a4f-a750-899b68adb4de-kube-api-access-jb979\") on node \"crc\" DevicePath \"\"" Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.419851 4944 generic.go:334] "Generic (PLEG): container finished" podID="d6a42739-d3f6-4a4f-a750-899b68adb4de" containerID="27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717" exitCode=0 Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.419938 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t882v" event={"ID":"d6a42739-d3f6-4a4f-a750-899b68adb4de","Type":"ContainerDied","Data":"27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717"} Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.419993 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t882v" event={"ID":"d6a42739-d3f6-4a4f-a750-899b68adb4de","Type":"ContainerDied","Data":"aacd727f30e4508a4e81c9480a51a6e679ff08f325bed48f161048384e179830"} Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.420017 4944 scope.go:117] "RemoveContainer" containerID="27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717" Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.420187 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t882v" Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.450287 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t882v"] Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.451953 4944 scope.go:117] "RemoveContainer" containerID="eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1" Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.457124 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t882v"] Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.474165 4944 scope.go:117] "RemoveContainer" containerID="9321673dbdb7d203b8f727fd8d077c524b7e30bfe87044a103c4fb6798b6f6f7" Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.474230 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.507431 4944 scope.go:117] "RemoveContainer" containerID="27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717" Nov 24 09:48:40 crc kubenswrapper[4944]: E1124 09:48:40.508015 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717\": container with ID starting with 27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717 not found: ID does not exist" containerID="27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717" Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.508061 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717"} err="failed to get container status \"27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717\": rpc error: code = NotFound desc = could not find container \"27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717\": container with ID starting with 27793c8e1e0c50160486dad66648e3813261127933dece087ecd1fde741ff717 not found: ID does not exist" Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.508082 4944 scope.go:117] "RemoveContainer" containerID="eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1" Nov 24 09:48:40 crc kubenswrapper[4944]: E1124 09:48:40.508507 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1\": container with ID starting with eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1 not found: ID does not exist" containerID="eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1" Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.508562 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1"} err="failed to get container status \"eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1\": rpc error: code = NotFound desc = could not find container \"eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1\": container with ID starting with eb42577db0013dd2776cbf3545648bd4ccd04453686dc499fd7760f6abfc63a1 not found: ID does not exist" Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.508593 4944 scope.go:117] "RemoveContainer" containerID="9321673dbdb7d203b8f727fd8d077c524b7e30bfe87044a103c4fb6798b6f6f7" Nov 24 09:48:40 crc kubenswrapper[4944]: E1124 09:48:40.511269 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9321673dbdb7d203b8f727fd8d077c524b7e30bfe87044a103c4fb6798b6f6f7\": container with ID starting with 9321673dbdb7d203b8f727fd8d077c524b7e30bfe87044a103c4fb6798b6f6f7 not found: ID does not exist" containerID="9321673dbdb7d203b8f727fd8d077c524b7e30bfe87044a103c4fb6798b6f6f7" Nov 24 09:48:40 crc kubenswrapper[4944]: I1124 09:48:40.511310 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9321673dbdb7d203b8f727fd8d077c524b7e30bfe87044a103c4fb6798b6f6f7"} err="failed to get container status \"9321673dbdb7d203b8f727fd8d077c524b7e30bfe87044a103c4fb6798b6f6f7\": rpc error: code = NotFound desc = could not find container \"9321673dbdb7d203b8f727fd8d077c524b7e30bfe87044a103c4fb6798b6f6f7\": container with ID starting with 9321673dbdb7d203b8f727fd8d077c524b7e30bfe87044a103c4fb6798b6f6f7 not found: ID does not exist" Nov 24 09:48:41 crc kubenswrapper[4944]: I1124 09:48:41.884378 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wf2rp"] Nov 24 09:48:42 crc kubenswrapper[4944]: I1124 09:48:42.287403 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6a42739-d3f6-4a4f-a750-899b68adb4de" path="/var/lib/kubelet/pods/d6a42739-d3f6-4a4f-a750-899b68adb4de/volumes" Nov 24 09:48:43 crc kubenswrapper[4944]: I1124 09:48:43.440847 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wf2rp" podUID="e92bfd19-27bd-413c-a718-c1279aff7dff" containerName="registry-server" containerID="cri-o://a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9" gracePeriod=2 Nov 24 09:48:43 crc kubenswrapper[4944]: I1124 09:48:43.869329 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:43 crc kubenswrapper[4944]: I1124 09:48:43.944367 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwks7\" (UniqueName: \"kubernetes.io/projected/e92bfd19-27bd-413c-a718-c1279aff7dff-kube-api-access-vwks7\") pod \"e92bfd19-27bd-413c-a718-c1279aff7dff\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " Nov 24 09:48:43 crc kubenswrapper[4944]: I1124 09:48:43.944445 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-utilities\") pod \"e92bfd19-27bd-413c-a718-c1279aff7dff\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " Nov 24 09:48:43 crc kubenswrapper[4944]: I1124 09:48:43.944552 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-catalog-content\") pod \"e92bfd19-27bd-413c-a718-c1279aff7dff\" (UID: \"e92bfd19-27bd-413c-a718-c1279aff7dff\") " Nov 24 09:48:43 crc kubenswrapper[4944]: I1124 09:48:43.945774 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-utilities" (OuterVolumeSpecName: "utilities") pod "e92bfd19-27bd-413c-a718-c1279aff7dff" (UID: "e92bfd19-27bd-413c-a718-c1279aff7dff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:48:43 crc kubenswrapper[4944]: I1124 09:48:43.949070 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e92bfd19-27bd-413c-a718-c1279aff7dff-kube-api-access-vwks7" (OuterVolumeSpecName: "kube-api-access-vwks7") pod "e92bfd19-27bd-413c-a718-c1279aff7dff" (UID: "e92bfd19-27bd-413c-a718-c1279aff7dff"). InnerVolumeSpecName "kube-api-access-vwks7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.035797 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e92bfd19-27bd-413c-a718-c1279aff7dff" (UID: "e92bfd19-27bd-413c-a718-c1279aff7dff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.046446 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.046488 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwks7\" (UniqueName: \"kubernetes.io/projected/e92bfd19-27bd-413c-a718-c1279aff7dff-kube-api-access-vwks7\") on node \"crc\" DevicePath \"\"" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.046509 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e92bfd19-27bd-413c-a718-c1279aff7dff-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.449898 4944 generic.go:334] "Generic (PLEG): container finished" podID="e92bfd19-27bd-413c-a718-c1279aff7dff" containerID="a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9" exitCode=0 Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.449939 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wf2rp" event={"ID":"e92bfd19-27bd-413c-a718-c1279aff7dff","Type":"ContainerDied","Data":"a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9"} Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.449966 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wf2rp" event={"ID":"e92bfd19-27bd-413c-a718-c1279aff7dff","Type":"ContainerDied","Data":"a1fee63dc1bab3654d542f8b38fe702d9f45590269d7d165b16e0c4fac2e6de2"} Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.449968 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wf2rp" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.449983 4944 scope.go:117] "RemoveContainer" containerID="a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.471929 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wf2rp"] Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.476843 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wf2rp"] Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.477787 4944 scope.go:117] "RemoveContainer" containerID="b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.493539 4944 scope.go:117] "RemoveContainer" containerID="39acb028e22fa3e871226569a8ac103505af8919205b80646de63c9904346a66" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.517092 4944 scope.go:117] "RemoveContainer" containerID="a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9" Nov 24 09:48:44 crc kubenswrapper[4944]: E1124 09:48:44.517497 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9\": container with ID starting with a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9 not found: ID does not exist" containerID="a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.517524 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9"} err="failed to get container status \"a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9\": rpc error: code = NotFound desc = could not find container \"a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9\": container with ID starting with a2606994971f3ae7fdce7b78e23b5dea293e8450e6cee6684d06b0dc493b2ac9 not found: ID does not exist" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.517543 4944 scope.go:117] "RemoveContainer" containerID="b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf" Nov 24 09:48:44 crc kubenswrapper[4944]: E1124 09:48:44.517917 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf\": container with ID starting with b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf not found: ID does not exist" containerID="b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.517956 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf"} err="failed to get container status \"b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf\": rpc error: code = NotFound desc = could not find container \"b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf\": container with ID starting with b5a3be79196234173883494b1e89ff9ebf9a7ed0e0cbffa52d8e144647711adf not found: ID does not exist" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.517987 4944 scope.go:117] "RemoveContainer" containerID="39acb028e22fa3e871226569a8ac103505af8919205b80646de63c9904346a66" Nov 24 09:48:44 crc kubenswrapper[4944]: E1124 09:48:44.518373 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39acb028e22fa3e871226569a8ac103505af8919205b80646de63c9904346a66\": container with ID starting with 39acb028e22fa3e871226569a8ac103505af8919205b80646de63c9904346a66 not found: ID does not exist" containerID="39acb028e22fa3e871226569a8ac103505af8919205b80646de63c9904346a66" Nov 24 09:48:44 crc kubenswrapper[4944]: I1124 09:48:44.518417 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39acb028e22fa3e871226569a8ac103505af8919205b80646de63c9904346a66"} err="failed to get container status \"39acb028e22fa3e871226569a8ac103505af8919205b80646de63c9904346a66\": rpc error: code = NotFound desc = could not find container \"39acb028e22fa3e871226569a8ac103505af8919205b80646de63c9904346a66\": container with ID starting with 39acb028e22fa3e871226569a8ac103505af8919205b80646de63c9904346a66 not found: ID does not exist" Nov 24 09:48:46 crc kubenswrapper[4944]: I1124 09:48:46.286293 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e92bfd19-27bd-413c-a718-c1279aff7dff" path="/var/lib/kubelet/pods/e92bfd19-27bd-413c-a718-c1279aff7dff/volumes" Nov 24 09:48:53 crc kubenswrapper[4944]: I1124 09:48:53.548503 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:48:53 crc kubenswrapper[4944]: I1124 09:48:53.549015 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.620074 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k8d7z"] Nov 24 09:49:12 crc kubenswrapper[4944]: E1124 09:49:12.620995 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a42739-d3f6-4a4f-a750-899b68adb4de" containerName="extract-utilities" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.621010 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a42739-d3f6-4a4f-a750-899b68adb4de" containerName="extract-utilities" Nov 24 09:49:12 crc kubenswrapper[4944]: E1124 09:49:12.621031 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e92bfd19-27bd-413c-a718-c1279aff7dff" containerName="extract-utilities" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.621039 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e92bfd19-27bd-413c-a718-c1279aff7dff" containerName="extract-utilities" Nov 24 09:49:12 crc kubenswrapper[4944]: E1124 09:49:12.621068 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a42739-d3f6-4a4f-a750-899b68adb4de" containerName="extract-content" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.621077 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a42739-d3f6-4a4f-a750-899b68adb4de" containerName="extract-content" Nov 24 09:49:12 crc kubenswrapper[4944]: E1124 09:49:12.621093 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e92bfd19-27bd-413c-a718-c1279aff7dff" containerName="extract-content" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.621100 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e92bfd19-27bd-413c-a718-c1279aff7dff" containerName="extract-content" Nov 24 09:49:12 crc kubenswrapper[4944]: E1124 09:49:12.621114 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a42739-d3f6-4a4f-a750-899b68adb4de" containerName="registry-server" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.621121 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a42739-d3f6-4a4f-a750-899b68adb4de" containerName="registry-server" Nov 24 09:49:12 crc kubenswrapper[4944]: E1124 09:49:12.621136 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e92bfd19-27bd-413c-a718-c1279aff7dff" containerName="registry-server" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.621142 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e92bfd19-27bd-413c-a718-c1279aff7dff" containerName="registry-server" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.621295 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a42739-d3f6-4a4f-a750-899b68adb4de" containerName="registry-server" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.621312 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e92bfd19-27bd-413c-a718-c1279aff7dff" containerName="registry-server" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.622840 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.638226 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8d7z"] Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.770958 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-utilities\") pod \"redhat-marketplace-k8d7z\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.771006 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-catalog-content\") pod \"redhat-marketplace-k8d7z\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.771127 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfhnc\" (UniqueName: \"kubernetes.io/projected/41af19cc-1980-4b7d-984e-df3a92271042-kube-api-access-hfhnc\") pod \"redhat-marketplace-k8d7z\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.872625 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfhnc\" (UniqueName: \"kubernetes.io/projected/41af19cc-1980-4b7d-984e-df3a92271042-kube-api-access-hfhnc\") pod \"redhat-marketplace-k8d7z\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.872735 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-utilities\") pod \"redhat-marketplace-k8d7z\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.872755 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-catalog-content\") pod \"redhat-marketplace-k8d7z\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.873355 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-utilities\") pod \"redhat-marketplace-k8d7z\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.873400 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-catalog-content\") pod \"redhat-marketplace-k8d7z\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.891748 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfhnc\" (UniqueName: \"kubernetes.io/projected/41af19cc-1980-4b7d-984e-df3a92271042-kube-api-access-hfhnc\") pod \"redhat-marketplace-k8d7z\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:12 crc kubenswrapper[4944]: I1124 09:49:12.940383 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:13 crc kubenswrapper[4944]: I1124 09:49:13.375017 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8d7z"] Nov 24 09:49:13 crc kubenswrapper[4944]: I1124 09:49:13.645487 4944 generic.go:334] "Generic (PLEG): container finished" podID="41af19cc-1980-4b7d-984e-df3a92271042" containerID="fca105750bdfd7a09c4d34598f9d0a4cefd93b0433cfda7954b2cae0e34770c5" exitCode=0 Nov 24 09:49:13 crc kubenswrapper[4944]: I1124 09:49:13.645534 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8d7z" event={"ID":"41af19cc-1980-4b7d-984e-df3a92271042","Type":"ContainerDied","Data":"fca105750bdfd7a09c4d34598f9d0a4cefd93b0433cfda7954b2cae0e34770c5"} Nov 24 09:49:13 crc kubenswrapper[4944]: I1124 09:49:13.645561 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8d7z" event={"ID":"41af19cc-1980-4b7d-984e-df3a92271042","Type":"ContainerStarted","Data":"b52a2d8c77a3b5b6c67e56ae08f79899faa87f7d94b5d29b983ac20f713dbd21"} Nov 24 09:49:14 crc kubenswrapper[4944]: I1124 09:49:14.657035 4944 generic.go:334] "Generic (PLEG): container finished" podID="41af19cc-1980-4b7d-984e-df3a92271042" containerID="98bfb6202901609343da48f2a7a42627217ef6e1568a4711994f2d115ca8eacc" exitCode=0 Nov 24 09:49:14 crc kubenswrapper[4944]: I1124 09:49:14.657112 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8d7z" event={"ID":"41af19cc-1980-4b7d-984e-df3a92271042","Type":"ContainerDied","Data":"98bfb6202901609343da48f2a7a42627217ef6e1568a4711994f2d115ca8eacc"} Nov 24 09:49:15 crc kubenswrapper[4944]: I1124 09:49:15.666270 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8d7z" event={"ID":"41af19cc-1980-4b7d-984e-df3a92271042","Type":"ContainerStarted","Data":"92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101"} Nov 24 09:49:15 crc kubenswrapper[4944]: I1124 09:49:15.682129 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k8d7z" podStartSLOduration=2.259782263 podStartE2EDuration="3.682112812s" podCreationTimestamp="2025-11-24 09:49:12 +0000 UTC" firstStartedPulling="2025-11-24 09:49:13.646851475 +0000 UTC m=+3414.181291937" lastFinishedPulling="2025-11-24 09:49:15.069182024 +0000 UTC m=+3415.603622486" observedRunningTime="2025-11-24 09:49:15.681036338 +0000 UTC m=+3416.215476800" watchObservedRunningTime="2025-11-24 09:49:15.682112812 +0000 UTC m=+3416.216553274" Nov 24 09:49:22 crc kubenswrapper[4944]: I1124 09:49:22.941510 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:22 crc kubenswrapper[4944]: I1124 09:49:22.943130 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:23 crc kubenswrapper[4944]: I1124 09:49:23.000734 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:23 crc kubenswrapper[4944]: I1124 09:49:23.548666 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:49:23 crc kubenswrapper[4944]: I1124 09:49:23.548747 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:49:23 crc kubenswrapper[4944]: I1124 09:49:23.548815 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:49:23 crc kubenswrapper[4944]: I1124 09:49:23.549471 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"de0b9680576b7802499acfa8e09920ec83923c1502e8d607ed1ab1120765d7ce"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:49:23 crc kubenswrapper[4944]: I1124 09:49:23.549531 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://de0b9680576b7802499acfa8e09920ec83923c1502e8d607ed1ab1120765d7ce" gracePeriod=600 Nov 24 09:49:23 crc kubenswrapper[4944]: I1124 09:49:23.728764 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="de0b9680576b7802499acfa8e09920ec83923c1502e8d607ed1ab1120765d7ce" exitCode=0 Nov 24 09:49:23 crc kubenswrapper[4944]: I1124 09:49:23.729413 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"de0b9680576b7802499acfa8e09920ec83923c1502e8d607ed1ab1120765d7ce"} Nov 24 09:49:23 crc kubenswrapper[4944]: I1124 09:49:23.729464 4944 scope.go:117] "RemoveContainer" containerID="0582c605991eb2b457e332afd3cbb569b88b8161459cd38dd88e51e7e151a859" Nov 24 09:49:23 crc kubenswrapper[4944]: I1124 09:49:23.783881 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:23 crc kubenswrapper[4944]: I1124 09:49:23.832947 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8d7z"] Nov 24 09:49:24 crc kubenswrapper[4944]: I1124 09:49:24.738143 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe"} Nov 24 09:49:25 crc kubenswrapper[4944]: I1124 09:49:25.744494 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k8d7z" podUID="41af19cc-1980-4b7d-984e-df3a92271042" containerName="registry-server" containerID="cri-o://92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101" gracePeriod=2 Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.117357 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.163756 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-utilities\") pod \"41af19cc-1980-4b7d-984e-df3a92271042\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.164465 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfhnc\" (UniqueName: \"kubernetes.io/projected/41af19cc-1980-4b7d-984e-df3a92271042-kube-api-access-hfhnc\") pod \"41af19cc-1980-4b7d-984e-df3a92271042\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.164495 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-catalog-content\") pod \"41af19cc-1980-4b7d-984e-df3a92271042\" (UID: \"41af19cc-1980-4b7d-984e-df3a92271042\") " Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.165471 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-utilities" (OuterVolumeSpecName: "utilities") pod "41af19cc-1980-4b7d-984e-df3a92271042" (UID: "41af19cc-1980-4b7d-984e-df3a92271042"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.171324 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41af19cc-1980-4b7d-984e-df3a92271042-kube-api-access-hfhnc" (OuterVolumeSpecName: "kube-api-access-hfhnc") pod "41af19cc-1980-4b7d-984e-df3a92271042" (UID: "41af19cc-1980-4b7d-984e-df3a92271042"). InnerVolumeSpecName "kube-api-access-hfhnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.189413 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41af19cc-1980-4b7d-984e-df3a92271042" (UID: "41af19cc-1980-4b7d-984e-df3a92271042"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.266761 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.266808 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfhnc\" (UniqueName: \"kubernetes.io/projected/41af19cc-1980-4b7d-984e-df3a92271042-kube-api-access-hfhnc\") on node \"crc\" DevicePath \"\"" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.266826 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41af19cc-1980-4b7d-984e-df3a92271042-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.752136 4944 generic.go:334] "Generic (PLEG): container finished" podID="41af19cc-1980-4b7d-984e-df3a92271042" containerID="92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101" exitCode=0 Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.752188 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k8d7z" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.752204 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8d7z" event={"ID":"41af19cc-1980-4b7d-984e-df3a92271042","Type":"ContainerDied","Data":"92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101"} Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.752587 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k8d7z" event={"ID":"41af19cc-1980-4b7d-984e-df3a92271042","Type":"ContainerDied","Data":"b52a2d8c77a3b5b6c67e56ae08f79899faa87f7d94b5d29b983ac20f713dbd21"} Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.752606 4944 scope.go:117] "RemoveContainer" containerID="92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.782247 4944 scope.go:117] "RemoveContainer" containerID="98bfb6202901609343da48f2a7a42627217ef6e1568a4711994f2d115ca8eacc" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.793816 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8d7z"] Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.804158 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k8d7z"] Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.808105 4944 scope.go:117] "RemoveContainer" containerID="fca105750bdfd7a09c4d34598f9d0a4cefd93b0433cfda7954b2cae0e34770c5" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.829408 4944 scope.go:117] "RemoveContainer" containerID="92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101" Nov 24 09:49:26 crc kubenswrapper[4944]: E1124 09:49:26.829915 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101\": container with ID starting with 92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101 not found: ID does not exist" containerID="92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.829991 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101"} err="failed to get container status \"92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101\": rpc error: code = NotFound desc = could not find container \"92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101\": container with ID starting with 92c3f10517ddd517e9e9c7e8c927beb28f4ae8d4bf851e8db8847510360e0101 not found: ID does not exist" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.830062 4944 scope.go:117] "RemoveContainer" containerID="98bfb6202901609343da48f2a7a42627217ef6e1568a4711994f2d115ca8eacc" Nov 24 09:49:26 crc kubenswrapper[4944]: E1124 09:49:26.830536 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98bfb6202901609343da48f2a7a42627217ef6e1568a4711994f2d115ca8eacc\": container with ID starting with 98bfb6202901609343da48f2a7a42627217ef6e1568a4711994f2d115ca8eacc not found: ID does not exist" containerID="98bfb6202901609343da48f2a7a42627217ef6e1568a4711994f2d115ca8eacc" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.830601 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98bfb6202901609343da48f2a7a42627217ef6e1568a4711994f2d115ca8eacc"} err="failed to get container status \"98bfb6202901609343da48f2a7a42627217ef6e1568a4711994f2d115ca8eacc\": rpc error: code = NotFound desc = could not find container \"98bfb6202901609343da48f2a7a42627217ef6e1568a4711994f2d115ca8eacc\": container with ID starting with 98bfb6202901609343da48f2a7a42627217ef6e1568a4711994f2d115ca8eacc not found: ID does not exist" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.830636 4944 scope.go:117] "RemoveContainer" containerID="fca105750bdfd7a09c4d34598f9d0a4cefd93b0433cfda7954b2cae0e34770c5" Nov 24 09:49:26 crc kubenswrapper[4944]: E1124 09:49:26.831481 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fca105750bdfd7a09c4d34598f9d0a4cefd93b0433cfda7954b2cae0e34770c5\": container with ID starting with fca105750bdfd7a09c4d34598f9d0a4cefd93b0433cfda7954b2cae0e34770c5 not found: ID does not exist" containerID="fca105750bdfd7a09c4d34598f9d0a4cefd93b0433cfda7954b2cae0e34770c5" Nov 24 09:49:26 crc kubenswrapper[4944]: I1124 09:49:26.831570 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fca105750bdfd7a09c4d34598f9d0a4cefd93b0433cfda7954b2cae0e34770c5"} err="failed to get container status \"fca105750bdfd7a09c4d34598f9d0a4cefd93b0433cfda7954b2cae0e34770c5\": rpc error: code = NotFound desc = could not find container \"fca105750bdfd7a09c4d34598f9d0a4cefd93b0433cfda7954b2cae0e34770c5\": container with ID starting with fca105750bdfd7a09c4d34598f9d0a4cefd93b0433cfda7954b2cae0e34770c5 not found: ID does not exist" Nov 24 09:49:28 crc kubenswrapper[4944]: I1124 09:49:28.289207 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41af19cc-1980-4b7d-984e-df3a92271042" path="/var/lib/kubelet/pods/41af19cc-1980-4b7d-984e-df3a92271042/volumes" Nov 24 09:50:50 crc kubenswrapper[4944]: I1124 09:50:50.970648 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zzqs2"] Nov 24 09:50:50 crc kubenswrapper[4944]: E1124 09:50:50.971592 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41af19cc-1980-4b7d-984e-df3a92271042" containerName="extract-utilities" Nov 24 09:50:50 crc kubenswrapper[4944]: I1124 09:50:50.971612 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="41af19cc-1980-4b7d-984e-df3a92271042" containerName="extract-utilities" Nov 24 09:50:50 crc kubenswrapper[4944]: E1124 09:50:50.971628 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41af19cc-1980-4b7d-984e-df3a92271042" containerName="registry-server" Nov 24 09:50:50 crc kubenswrapper[4944]: I1124 09:50:50.971636 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="41af19cc-1980-4b7d-984e-df3a92271042" containerName="registry-server" Nov 24 09:50:50 crc kubenswrapper[4944]: E1124 09:50:50.971677 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41af19cc-1980-4b7d-984e-df3a92271042" containerName="extract-content" Nov 24 09:50:50 crc kubenswrapper[4944]: I1124 09:50:50.971685 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="41af19cc-1980-4b7d-984e-df3a92271042" containerName="extract-content" Nov 24 09:50:50 crc kubenswrapper[4944]: I1124 09:50:50.971852 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="41af19cc-1980-4b7d-984e-df3a92271042" containerName="registry-server" Nov 24 09:50:50 crc kubenswrapper[4944]: I1124 09:50:50.975035 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:50:50 crc kubenswrapper[4944]: I1124 09:50:50.980447 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zzqs2"] Nov 24 09:50:51 crc kubenswrapper[4944]: I1124 09:50:51.150810 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npwxs\" (UniqueName: \"kubernetes.io/projected/04e07705-6d22-4751-b60e-255dad104ab1-kube-api-access-npwxs\") pod \"community-operators-zzqs2\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:50:51 crc kubenswrapper[4944]: I1124 09:50:51.150885 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-utilities\") pod \"community-operators-zzqs2\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:50:51 crc kubenswrapper[4944]: I1124 09:50:51.150946 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-catalog-content\") pod \"community-operators-zzqs2\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:50:51 crc kubenswrapper[4944]: I1124 09:50:51.251979 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npwxs\" (UniqueName: \"kubernetes.io/projected/04e07705-6d22-4751-b60e-255dad104ab1-kube-api-access-npwxs\") pod \"community-operators-zzqs2\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:50:51 crc kubenswrapper[4944]: I1124 09:50:51.252040 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-utilities\") pod \"community-operators-zzqs2\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:50:51 crc kubenswrapper[4944]: I1124 09:50:51.252078 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-catalog-content\") pod \"community-operators-zzqs2\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:50:51 crc kubenswrapper[4944]: I1124 09:50:51.252643 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-utilities\") pod \"community-operators-zzqs2\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:50:51 crc kubenswrapper[4944]: I1124 09:50:51.252688 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-catalog-content\") pod \"community-operators-zzqs2\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:50:51 crc kubenswrapper[4944]: I1124 09:50:51.272321 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npwxs\" (UniqueName: \"kubernetes.io/projected/04e07705-6d22-4751-b60e-255dad104ab1-kube-api-access-npwxs\") pod \"community-operators-zzqs2\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:50:51 crc kubenswrapper[4944]: I1124 09:50:51.294519 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:50:51 crc kubenswrapper[4944]: I1124 09:50:51.811704 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zzqs2"] Nov 24 09:50:51 crc kubenswrapper[4944]: W1124 09:50:51.815576 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04e07705_6d22_4751_b60e_255dad104ab1.slice/crio-ecaf600815b36cce5e6550b27c0c8ee09f30f0e6c3e5577b62ae9c8174f9c0f6 WatchSource:0}: Error finding container ecaf600815b36cce5e6550b27c0c8ee09f30f0e6c3e5577b62ae9c8174f9c0f6: Status 404 returned error can't find the container with id ecaf600815b36cce5e6550b27c0c8ee09f30f0e6c3e5577b62ae9c8174f9c0f6 Nov 24 09:50:52 crc kubenswrapper[4944]: I1124 09:50:52.355003 4944 generic.go:334] "Generic (PLEG): container finished" podID="04e07705-6d22-4751-b60e-255dad104ab1" containerID="cb3b5cf31d87e4b58cfa44f2b8ae166f836d58a4fa2349dc3901b1e73c7eddd5" exitCode=0 Nov 24 09:50:52 crc kubenswrapper[4944]: I1124 09:50:52.355112 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqs2" event={"ID":"04e07705-6d22-4751-b60e-255dad104ab1","Type":"ContainerDied","Data":"cb3b5cf31d87e4b58cfa44f2b8ae166f836d58a4fa2349dc3901b1e73c7eddd5"} Nov 24 09:50:52 crc kubenswrapper[4944]: I1124 09:50:52.355377 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqs2" event={"ID":"04e07705-6d22-4751-b60e-255dad104ab1","Type":"ContainerStarted","Data":"ecaf600815b36cce5e6550b27c0c8ee09f30f0e6c3e5577b62ae9c8174f9c0f6"} Nov 24 09:50:53 crc kubenswrapper[4944]: I1124 09:50:53.365101 4944 generic.go:334] "Generic (PLEG): container finished" podID="04e07705-6d22-4751-b60e-255dad104ab1" containerID="18eba5b3988c7a73a7779837a92e95367acf3affdc6d574b2593218a06ef4003" exitCode=0 Nov 24 09:50:53 crc kubenswrapper[4944]: I1124 09:50:53.365153 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqs2" event={"ID":"04e07705-6d22-4751-b60e-255dad104ab1","Type":"ContainerDied","Data":"18eba5b3988c7a73a7779837a92e95367acf3affdc6d574b2593218a06ef4003"} Nov 24 09:50:54 crc kubenswrapper[4944]: I1124 09:50:54.387819 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqs2" event={"ID":"04e07705-6d22-4751-b60e-255dad104ab1","Type":"ContainerStarted","Data":"cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75"} Nov 24 09:50:54 crc kubenswrapper[4944]: I1124 09:50:54.406093 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zzqs2" podStartSLOduration=2.997529644 podStartE2EDuration="4.406073045s" podCreationTimestamp="2025-11-24 09:50:50 +0000 UTC" firstStartedPulling="2025-11-24 09:50:52.356536402 +0000 UTC m=+3512.890976864" lastFinishedPulling="2025-11-24 09:50:53.765079803 +0000 UTC m=+3514.299520265" observedRunningTime="2025-11-24 09:50:54.404662159 +0000 UTC m=+3514.939102631" watchObservedRunningTime="2025-11-24 09:50:54.406073045 +0000 UTC m=+3514.940513507" Nov 24 09:51:01 crc kubenswrapper[4944]: I1124 09:51:01.295158 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:51:01 crc kubenswrapper[4944]: I1124 09:51:01.295776 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:51:01 crc kubenswrapper[4944]: I1124 09:51:01.339947 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:51:01 crc kubenswrapper[4944]: I1124 09:51:01.482761 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:51:01 crc kubenswrapper[4944]: I1124 09:51:01.569158 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zzqs2"] Nov 24 09:51:03 crc kubenswrapper[4944]: I1124 09:51:03.452172 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zzqs2" podUID="04e07705-6d22-4751-b60e-255dad104ab1" containerName="registry-server" containerID="cri-o://cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75" gracePeriod=2 Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.294854 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.448250 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-catalog-content\") pod \"04e07705-6d22-4751-b60e-255dad104ab1\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.448300 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-utilities\") pod \"04e07705-6d22-4751-b60e-255dad104ab1\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.448387 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npwxs\" (UniqueName: \"kubernetes.io/projected/04e07705-6d22-4751-b60e-255dad104ab1-kube-api-access-npwxs\") pod \"04e07705-6d22-4751-b60e-255dad104ab1\" (UID: \"04e07705-6d22-4751-b60e-255dad104ab1\") " Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.449411 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-utilities" (OuterVolumeSpecName: "utilities") pod "04e07705-6d22-4751-b60e-255dad104ab1" (UID: "04e07705-6d22-4751-b60e-255dad104ab1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.453873 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04e07705-6d22-4751-b60e-255dad104ab1-kube-api-access-npwxs" (OuterVolumeSpecName: "kube-api-access-npwxs") pod "04e07705-6d22-4751-b60e-255dad104ab1" (UID: "04e07705-6d22-4751-b60e-255dad104ab1"). InnerVolumeSpecName "kube-api-access-npwxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.461594 4944 generic.go:334] "Generic (PLEG): container finished" podID="04e07705-6d22-4751-b60e-255dad104ab1" containerID="cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75" exitCode=0 Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.461630 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zzqs2" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.461639 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqs2" event={"ID":"04e07705-6d22-4751-b60e-255dad104ab1","Type":"ContainerDied","Data":"cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75"} Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.461668 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zzqs2" event={"ID":"04e07705-6d22-4751-b60e-255dad104ab1","Type":"ContainerDied","Data":"ecaf600815b36cce5e6550b27c0c8ee09f30f0e6c3e5577b62ae9c8174f9c0f6"} Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.461687 4944 scope.go:117] "RemoveContainer" containerID="cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.489074 4944 scope.go:117] "RemoveContainer" containerID="18eba5b3988c7a73a7779837a92e95367acf3affdc6d574b2593218a06ef4003" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.502122 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04e07705-6d22-4751-b60e-255dad104ab1" (UID: "04e07705-6d22-4751-b60e-255dad104ab1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.504862 4944 scope.go:117] "RemoveContainer" containerID="cb3b5cf31d87e4b58cfa44f2b8ae166f836d58a4fa2349dc3901b1e73c7eddd5" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.529448 4944 scope.go:117] "RemoveContainer" containerID="cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75" Nov 24 09:51:04 crc kubenswrapper[4944]: E1124 09:51:04.529878 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75\": container with ID starting with cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75 not found: ID does not exist" containerID="cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.529922 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75"} err="failed to get container status \"cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75\": rpc error: code = NotFound desc = could not find container \"cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75\": container with ID starting with cb7b3af3e0354d697cb4d60fadaaec340c182467e958a9e6c9ae194f837bac75 not found: ID does not exist" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.529948 4944 scope.go:117] "RemoveContainer" containerID="18eba5b3988c7a73a7779837a92e95367acf3affdc6d574b2593218a06ef4003" Nov 24 09:51:04 crc kubenswrapper[4944]: E1124 09:51:04.530426 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18eba5b3988c7a73a7779837a92e95367acf3affdc6d574b2593218a06ef4003\": container with ID starting with 18eba5b3988c7a73a7779837a92e95367acf3affdc6d574b2593218a06ef4003 not found: ID does not exist" containerID="18eba5b3988c7a73a7779837a92e95367acf3affdc6d574b2593218a06ef4003" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.530455 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18eba5b3988c7a73a7779837a92e95367acf3affdc6d574b2593218a06ef4003"} err="failed to get container status \"18eba5b3988c7a73a7779837a92e95367acf3affdc6d574b2593218a06ef4003\": rpc error: code = NotFound desc = could not find container \"18eba5b3988c7a73a7779837a92e95367acf3affdc6d574b2593218a06ef4003\": container with ID starting with 18eba5b3988c7a73a7779837a92e95367acf3affdc6d574b2593218a06ef4003 not found: ID does not exist" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.530476 4944 scope.go:117] "RemoveContainer" containerID="cb3b5cf31d87e4b58cfa44f2b8ae166f836d58a4fa2349dc3901b1e73c7eddd5" Nov 24 09:51:04 crc kubenswrapper[4944]: E1124 09:51:04.530748 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb3b5cf31d87e4b58cfa44f2b8ae166f836d58a4fa2349dc3901b1e73c7eddd5\": container with ID starting with cb3b5cf31d87e4b58cfa44f2b8ae166f836d58a4fa2349dc3901b1e73c7eddd5 not found: ID does not exist" containerID="cb3b5cf31d87e4b58cfa44f2b8ae166f836d58a4fa2349dc3901b1e73c7eddd5" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.530769 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb3b5cf31d87e4b58cfa44f2b8ae166f836d58a4fa2349dc3901b1e73c7eddd5"} err="failed to get container status \"cb3b5cf31d87e4b58cfa44f2b8ae166f836d58a4fa2349dc3901b1e73c7eddd5\": rpc error: code = NotFound desc = could not find container \"cb3b5cf31d87e4b58cfa44f2b8ae166f836d58a4fa2349dc3901b1e73c7eddd5\": container with ID starting with cb3b5cf31d87e4b58cfa44f2b8ae166f836d58a4fa2349dc3901b1e73c7eddd5 not found: ID does not exist" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.549564 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npwxs\" (UniqueName: \"kubernetes.io/projected/04e07705-6d22-4751-b60e-255dad104ab1-kube-api-access-npwxs\") on node \"crc\" DevicePath \"\"" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.549592 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.549602 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04e07705-6d22-4751-b60e-255dad104ab1-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.794236 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zzqs2"] Nov 24 09:51:04 crc kubenswrapper[4944]: I1124 09:51:04.799803 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zzqs2"] Nov 24 09:51:06 crc kubenswrapper[4944]: I1124 09:51:06.291730 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04e07705-6d22-4751-b60e-255dad104ab1" path="/var/lib/kubelet/pods/04e07705-6d22-4751-b60e-255dad104ab1/volumes" Nov 24 09:51:23 crc kubenswrapper[4944]: I1124 09:51:23.548893 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:51:23 crc kubenswrapper[4944]: I1124 09:51:23.549568 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:51:53 crc kubenswrapper[4944]: I1124 09:51:53.548758 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:51:53 crc kubenswrapper[4944]: I1124 09:51:53.550615 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:52:23 crc kubenswrapper[4944]: I1124 09:52:23.548598 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:52:23 crc kubenswrapper[4944]: I1124 09:52:23.549223 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:52:23 crc kubenswrapper[4944]: I1124 09:52:23.549273 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 09:52:23 crc kubenswrapper[4944]: I1124 09:52:23.550011 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:52:23 crc kubenswrapper[4944]: I1124 09:52:23.550094 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" gracePeriod=600 Nov 24 09:52:23 crc kubenswrapper[4944]: E1124 09:52:23.669911 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:52:24 crc kubenswrapper[4944]: I1124 09:52:24.019529 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" exitCode=0 Nov 24 09:52:24 crc kubenswrapper[4944]: I1124 09:52:24.019600 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe"} Nov 24 09:52:24 crc kubenswrapper[4944]: I1124 09:52:24.019899 4944 scope.go:117] "RemoveContainer" containerID="de0b9680576b7802499acfa8e09920ec83923c1502e8d607ed1ab1120765d7ce" Nov 24 09:52:24 crc kubenswrapper[4944]: I1124 09:52:24.020929 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:52:24 crc kubenswrapper[4944]: E1124 09:52:24.021708 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:52:35 crc kubenswrapper[4944]: I1124 09:52:35.277014 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:52:35 crc kubenswrapper[4944]: E1124 09:52:35.278331 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:52:50 crc kubenswrapper[4944]: I1124 09:52:50.282096 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:52:50 crc kubenswrapper[4944]: E1124 09:52:50.282890 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:53:03 crc kubenswrapper[4944]: I1124 09:53:03.277976 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:53:03 crc kubenswrapper[4944]: E1124 09:53:03.279041 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:53:16 crc kubenswrapper[4944]: I1124 09:53:16.276879 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:53:16 crc kubenswrapper[4944]: E1124 09:53:16.278071 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:53:27 crc kubenswrapper[4944]: I1124 09:53:27.277202 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:53:27 crc kubenswrapper[4944]: E1124 09:53:27.277896 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:53:41 crc kubenswrapper[4944]: I1124 09:53:41.277039 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:53:41 crc kubenswrapper[4944]: E1124 09:53:41.277857 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:53:56 crc kubenswrapper[4944]: I1124 09:53:56.276518 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:53:56 crc kubenswrapper[4944]: E1124 09:53:56.277257 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:54:10 crc kubenswrapper[4944]: I1124 09:54:10.283231 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:54:10 crc kubenswrapper[4944]: E1124 09:54:10.284573 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:54:24 crc kubenswrapper[4944]: I1124 09:54:24.276904 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:54:24 crc kubenswrapper[4944]: E1124 09:54:24.277484 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:54:36 crc kubenswrapper[4944]: I1124 09:54:36.276973 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:54:36 crc kubenswrapper[4944]: E1124 09:54:36.277785 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:54:47 crc kubenswrapper[4944]: I1124 09:54:47.276824 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:54:47 crc kubenswrapper[4944]: E1124 09:54:47.277578 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:55:00 crc kubenswrapper[4944]: I1124 09:55:00.280792 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:55:00 crc kubenswrapper[4944]: E1124 09:55:00.281617 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:55:15 crc kubenswrapper[4944]: I1124 09:55:15.277377 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:55:15 crc kubenswrapper[4944]: E1124 09:55:15.278095 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:55:28 crc kubenswrapper[4944]: I1124 09:55:28.276954 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:55:28 crc kubenswrapper[4944]: E1124 09:55:28.277675 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:55:39 crc kubenswrapper[4944]: I1124 09:55:39.276985 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:55:39 crc kubenswrapper[4944]: E1124 09:55:39.278361 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:55:50 crc kubenswrapper[4944]: I1124 09:55:50.282022 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:55:50 crc kubenswrapper[4944]: E1124 09:55:50.283982 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:56:05 crc kubenswrapper[4944]: I1124 09:56:05.276649 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:56:05 crc kubenswrapper[4944]: E1124 09:56:05.278341 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:56:20 crc kubenswrapper[4944]: I1124 09:56:20.280706 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:56:20 crc kubenswrapper[4944]: E1124 09:56:20.282858 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:56:32 crc kubenswrapper[4944]: I1124 09:56:32.276205 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:56:32 crc kubenswrapper[4944]: E1124 09:56:32.276877 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:56:45 crc kubenswrapper[4944]: I1124 09:56:45.277438 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:56:45 crc kubenswrapper[4944]: E1124 09:56:45.278377 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:56:57 crc kubenswrapper[4944]: I1124 09:56:57.277687 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:56:57 crc kubenswrapper[4944]: E1124 09:56:57.278900 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:57:09 crc kubenswrapper[4944]: I1124 09:57:09.277898 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:57:09 crc kubenswrapper[4944]: E1124 09:57:09.278699 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:57:23 crc kubenswrapper[4944]: I1124 09:57:23.276854 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:57:23 crc kubenswrapper[4944]: E1124 09:57:23.277502 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 09:57:38 crc kubenswrapper[4944]: I1124 09:57:38.277133 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 09:57:39 crc kubenswrapper[4944]: I1124 09:57:39.170074 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"e5cc635a58a05c32265a03514792334cbb1f6a83a754bd0ae32e207ab72d886d"} Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.677744 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-94x4w"] Nov 24 09:59:02 crc kubenswrapper[4944]: E1124 09:59:02.678852 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04e07705-6d22-4751-b60e-255dad104ab1" containerName="extract-utilities" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.678866 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="04e07705-6d22-4751-b60e-255dad104ab1" containerName="extract-utilities" Nov 24 09:59:02 crc kubenswrapper[4944]: E1124 09:59:02.678883 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04e07705-6d22-4751-b60e-255dad104ab1" containerName="extract-content" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.678889 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="04e07705-6d22-4751-b60e-255dad104ab1" containerName="extract-content" Nov 24 09:59:02 crc kubenswrapper[4944]: E1124 09:59:02.678916 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04e07705-6d22-4751-b60e-255dad104ab1" containerName="registry-server" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.678922 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="04e07705-6d22-4751-b60e-255dad104ab1" containerName="registry-server" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.679090 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="04e07705-6d22-4751-b60e-255dad104ab1" containerName="registry-server" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.680358 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.697767 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-94x4w"] Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.843760 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-utilities\") pod \"redhat-operators-94x4w\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.843827 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxzsp\" (UniqueName: \"kubernetes.io/projected/7af46679-4e8e-4f62-978e-6bfbfe04f210-kube-api-access-gxzsp\") pod \"redhat-operators-94x4w\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.843854 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-catalog-content\") pod \"redhat-operators-94x4w\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.944939 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-utilities\") pod \"redhat-operators-94x4w\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.945009 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxzsp\" (UniqueName: \"kubernetes.io/projected/7af46679-4e8e-4f62-978e-6bfbfe04f210-kube-api-access-gxzsp\") pod \"redhat-operators-94x4w\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.945032 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-catalog-content\") pod \"redhat-operators-94x4w\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.945500 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-utilities\") pod \"redhat-operators-94x4w\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.945562 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-catalog-content\") pod \"redhat-operators-94x4w\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:02 crc kubenswrapper[4944]: I1124 09:59:02.966552 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxzsp\" (UniqueName: \"kubernetes.io/projected/7af46679-4e8e-4f62-978e-6bfbfe04f210-kube-api-access-gxzsp\") pod \"redhat-operators-94x4w\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:03 crc kubenswrapper[4944]: I1124 09:59:03.004546 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:03 crc kubenswrapper[4944]: I1124 09:59:03.430970 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-94x4w"] Nov 24 09:59:03 crc kubenswrapper[4944]: I1124 09:59:03.758802 4944 generic.go:334] "Generic (PLEG): container finished" podID="7af46679-4e8e-4f62-978e-6bfbfe04f210" containerID="68850ef2b6ad5d58f7c718b88c5bac14d74d3d38b9d23fcc3465af53b936cbd1" exitCode=0 Nov 24 09:59:03 crc kubenswrapper[4944]: I1124 09:59:03.758858 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94x4w" event={"ID":"7af46679-4e8e-4f62-978e-6bfbfe04f210","Type":"ContainerDied","Data":"68850ef2b6ad5d58f7c718b88c5bac14d74d3d38b9d23fcc3465af53b936cbd1"} Nov 24 09:59:03 crc kubenswrapper[4944]: I1124 09:59:03.759149 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94x4w" event={"ID":"7af46679-4e8e-4f62-978e-6bfbfe04f210","Type":"ContainerStarted","Data":"7c6395e06d90011a48b6433ba7e52d571d4c1a00b80e7db61104a7e5c0088bb7"} Nov 24 09:59:03 crc kubenswrapper[4944]: I1124 09:59:03.760941 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:59:05 crc kubenswrapper[4944]: I1124 09:59:05.774816 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94x4w" event={"ID":"7af46679-4e8e-4f62-978e-6bfbfe04f210","Type":"ContainerStarted","Data":"165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7"} Nov 24 09:59:06 crc kubenswrapper[4944]: I1124 09:59:06.782727 4944 generic.go:334] "Generic (PLEG): container finished" podID="7af46679-4e8e-4f62-978e-6bfbfe04f210" containerID="165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7" exitCode=0 Nov 24 09:59:06 crc kubenswrapper[4944]: I1124 09:59:06.782795 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94x4w" event={"ID":"7af46679-4e8e-4f62-978e-6bfbfe04f210","Type":"ContainerDied","Data":"165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7"} Nov 24 09:59:07 crc kubenswrapper[4944]: I1124 09:59:07.791732 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94x4w" event={"ID":"7af46679-4e8e-4f62-978e-6bfbfe04f210","Type":"ContainerStarted","Data":"768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618"} Nov 24 09:59:07 crc kubenswrapper[4944]: I1124 09:59:07.808361 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-94x4w" podStartSLOduration=2.336955567 podStartE2EDuration="5.808341934s" podCreationTimestamp="2025-11-24 09:59:02 +0000 UTC" firstStartedPulling="2025-11-24 09:59:03.760567902 +0000 UTC m=+4004.295008374" lastFinishedPulling="2025-11-24 09:59:07.231954279 +0000 UTC m=+4007.766394741" observedRunningTime="2025-11-24 09:59:07.806180876 +0000 UTC m=+4008.340621358" watchObservedRunningTime="2025-11-24 09:59:07.808341934 +0000 UTC m=+4008.342782396" Nov 24 09:59:13 crc kubenswrapper[4944]: I1124 09:59:13.005317 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:13 crc kubenswrapper[4944]: I1124 09:59:13.009610 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:13 crc kubenswrapper[4944]: I1124 09:59:13.050041 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:13 crc kubenswrapper[4944]: I1124 09:59:13.932302 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:13 crc kubenswrapper[4944]: I1124 09:59:13.975507 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-94x4w"] Nov 24 09:59:15 crc kubenswrapper[4944]: I1124 09:59:15.851802 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-94x4w" podUID="7af46679-4e8e-4f62-978e-6bfbfe04f210" containerName="registry-server" containerID="cri-o://768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618" gracePeriod=2 Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.373496 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.419078 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-utilities\") pod \"7af46679-4e8e-4f62-978e-6bfbfe04f210\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.419142 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxzsp\" (UniqueName: \"kubernetes.io/projected/7af46679-4e8e-4f62-978e-6bfbfe04f210-kube-api-access-gxzsp\") pod \"7af46679-4e8e-4f62-978e-6bfbfe04f210\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.419215 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-catalog-content\") pod \"7af46679-4e8e-4f62-978e-6bfbfe04f210\" (UID: \"7af46679-4e8e-4f62-978e-6bfbfe04f210\") " Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.420073 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-utilities" (OuterVolumeSpecName: "utilities") pod "7af46679-4e8e-4f62-978e-6bfbfe04f210" (UID: "7af46679-4e8e-4f62-978e-6bfbfe04f210"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.427325 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7af46679-4e8e-4f62-978e-6bfbfe04f210-kube-api-access-gxzsp" (OuterVolumeSpecName: "kube-api-access-gxzsp") pod "7af46679-4e8e-4f62-978e-6bfbfe04f210" (UID: "7af46679-4e8e-4f62-978e-6bfbfe04f210"). InnerVolumeSpecName "kube-api-access-gxzsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.520377 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxzsp\" (UniqueName: \"kubernetes.io/projected/7af46679-4e8e-4f62-978e-6bfbfe04f210-kube-api-access-gxzsp\") on node \"crc\" DevicePath \"\"" Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.520409 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.861888 4944 generic.go:334] "Generic (PLEG): container finished" podID="7af46679-4e8e-4f62-978e-6bfbfe04f210" containerID="768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618" exitCode=0 Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.861971 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94x4w" event={"ID":"7af46679-4e8e-4f62-978e-6bfbfe04f210","Type":"ContainerDied","Data":"768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618"} Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.862019 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-94x4w" event={"ID":"7af46679-4e8e-4f62-978e-6bfbfe04f210","Type":"ContainerDied","Data":"7c6395e06d90011a48b6433ba7e52d571d4c1a00b80e7db61104a7e5c0088bb7"} Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.862073 4944 scope.go:117] "RemoveContainer" containerID="768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618" Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.862320 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-94x4w" Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.882001 4944 scope.go:117] "RemoveContainer" containerID="165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7" Nov 24 09:59:16 crc kubenswrapper[4944]: I1124 09:59:16.897399 4944 scope.go:117] "RemoveContainer" containerID="68850ef2b6ad5d58f7c718b88c5bac14d74d3d38b9d23fcc3465af53b936cbd1" Nov 24 09:59:17 crc kubenswrapper[4944]: I1124 09:59:17.129564 4944 scope.go:117] "RemoveContainer" containerID="768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618" Nov 24 09:59:17 crc kubenswrapper[4944]: E1124 09:59:17.130088 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618\": container with ID starting with 768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618 not found: ID does not exist" containerID="768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618" Nov 24 09:59:17 crc kubenswrapper[4944]: I1124 09:59:17.130127 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618"} err="failed to get container status \"768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618\": rpc error: code = NotFound desc = could not find container \"768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618\": container with ID starting with 768c67a65fc224817659b9930a768a4b645efd47405a0532ebc12ea594e7c618 not found: ID does not exist" Nov 24 09:59:17 crc kubenswrapper[4944]: I1124 09:59:17.130152 4944 scope.go:117] "RemoveContainer" containerID="165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7" Nov 24 09:59:17 crc kubenswrapper[4944]: E1124 09:59:17.130544 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7\": container with ID starting with 165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7 not found: ID does not exist" containerID="165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7" Nov 24 09:59:17 crc kubenswrapper[4944]: I1124 09:59:17.130565 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7"} err="failed to get container status \"165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7\": rpc error: code = NotFound desc = could not find container \"165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7\": container with ID starting with 165d83ef36a35c56c684d279bf1ed7edc0618cf2a847b7fa4af45daef7f8dbf7 not found: ID does not exist" Nov 24 09:59:17 crc kubenswrapper[4944]: I1124 09:59:17.130580 4944 scope.go:117] "RemoveContainer" containerID="68850ef2b6ad5d58f7c718b88c5bac14d74d3d38b9d23fcc3465af53b936cbd1" Nov 24 09:59:17 crc kubenswrapper[4944]: E1124 09:59:17.130880 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68850ef2b6ad5d58f7c718b88c5bac14d74d3d38b9d23fcc3465af53b936cbd1\": container with ID starting with 68850ef2b6ad5d58f7c718b88c5bac14d74d3d38b9d23fcc3465af53b936cbd1 not found: ID does not exist" containerID="68850ef2b6ad5d58f7c718b88c5bac14d74d3d38b9d23fcc3465af53b936cbd1" Nov 24 09:59:17 crc kubenswrapper[4944]: I1124 09:59:17.130900 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68850ef2b6ad5d58f7c718b88c5bac14d74d3d38b9d23fcc3465af53b936cbd1"} err="failed to get container status \"68850ef2b6ad5d58f7c718b88c5bac14d74d3d38b9d23fcc3465af53b936cbd1\": rpc error: code = NotFound desc = could not find container \"68850ef2b6ad5d58f7c718b88c5bac14d74d3d38b9d23fcc3465af53b936cbd1\": container with ID starting with 68850ef2b6ad5d58f7c718b88c5bac14d74d3d38b9d23fcc3465af53b936cbd1 not found: ID does not exist" Nov 24 09:59:17 crc kubenswrapper[4944]: I1124 09:59:17.245008 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7af46679-4e8e-4f62-978e-6bfbfe04f210" (UID: "7af46679-4e8e-4f62-978e-6bfbfe04f210"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:59:17 crc kubenswrapper[4944]: I1124 09:59:17.332769 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7af46679-4e8e-4f62-978e-6bfbfe04f210-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:59:17 crc kubenswrapper[4944]: I1124 09:59:17.496905 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-94x4w"] Nov 24 09:59:17 crc kubenswrapper[4944]: I1124 09:59:17.501985 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-94x4w"] Nov 24 09:59:18 crc kubenswrapper[4944]: I1124 09:59:18.284400 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7af46679-4e8e-4f62-978e-6bfbfe04f210" path="/var/lib/kubelet/pods/7af46679-4e8e-4f62-978e-6bfbfe04f210/volumes" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.095754 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lx8k4"] Nov 24 09:59:41 crc kubenswrapper[4944]: E1124 09:59:41.096623 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af46679-4e8e-4f62-978e-6bfbfe04f210" containerName="registry-server" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.096636 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af46679-4e8e-4f62-978e-6bfbfe04f210" containerName="registry-server" Nov 24 09:59:41 crc kubenswrapper[4944]: E1124 09:59:41.096660 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af46679-4e8e-4f62-978e-6bfbfe04f210" containerName="extract-utilities" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.096666 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af46679-4e8e-4f62-978e-6bfbfe04f210" containerName="extract-utilities" Nov 24 09:59:41 crc kubenswrapper[4944]: E1124 09:59:41.096678 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af46679-4e8e-4f62-978e-6bfbfe04f210" containerName="extract-content" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.096684 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af46679-4e8e-4f62-978e-6bfbfe04f210" containerName="extract-content" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.096819 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="7af46679-4e8e-4f62-978e-6bfbfe04f210" containerName="registry-server" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.099440 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.106729 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lx8k4"] Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.180827 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-utilities\") pod \"certified-operators-lx8k4\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.180952 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9nv5\" (UniqueName: \"kubernetes.io/projected/759333d0-7fcf-4c0b-8733-083e9d5fcc09-kube-api-access-w9nv5\") pod \"certified-operators-lx8k4\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.180980 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-catalog-content\") pod \"certified-operators-lx8k4\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.282126 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9nv5\" (UniqueName: \"kubernetes.io/projected/759333d0-7fcf-4c0b-8733-083e9d5fcc09-kube-api-access-w9nv5\") pod \"certified-operators-lx8k4\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.282189 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-catalog-content\") pod \"certified-operators-lx8k4\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.282238 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-utilities\") pod \"certified-operators-lx8k4\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.282841 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-utilities\") pod \"certified-operators-lx8k4\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.282882 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-catalog-content\") pod \"certified-operators-lx8k4\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.303084 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9nv5\" (UniqueName: \"kubernetes.io/projected/759333d0-7fcf-4c0b-8733-083e9d5fcc09-kube-api-access-w9nv5\") pod \"certified-operators-lx8k4\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.421610 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:41 crc kubenswrapper[4944]: I1124 09:59:41.899838 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lx8k4"] Nov 24 09:59:42 crc kubenswrapper[4944]: I1124 09:59:42.069653 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx8k4" event={"ID":"759333d0-7fcf-4c0b-8733-083e9d5fcc09","Type":"ContainerStarted","Data":"a68d581240bab92a77d5dd620e3124aa56debeb655d52aa1a02a31449427cc89"} Nov 24 09:59:43 crc kubenswrapper[4944]: I1124 09:59:43.077740 4944 generic.go:334] "Generic (PLEG): container finished" podID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" containerID="db5e214823c58c0ff53d7bd41b8bef496d91481f5eb15a053b0ff99134d51a80" exitCode=0 Nov 24 09:59:43 crc kubenswrapper[4944]: I1124 09:59:43.077782 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx8k4" event={"ID":"759333d0-7fcf-4c0b-8733-083e9d5fcc09","Type":"ContainerDied","Data":"db5e214823c58c0ff53d7bd41b8bef496d91481f5eb15a053b0ff99134d51a80"} Nov 24 09:59:44 crc kubenswrapper[4944]: I1124 09:59:44.103552 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx8k4" event={"ID":"759333d0-7fcf-4c0b-8733-083e9d5fcc09","Type":"ContainerStarted","Data":"f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7"} Nov 24 09:59:45 crc kubenswrapper[4944]: I1124 09:59:45.113789 4944 generic.go:334] "Generic (PLEG): container finished" podID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" containerID="f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7" exitCode=0 Nov 24 09:59:45 crc kubenswrapper[4944]: I1124 09:59:45.113914 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx8k4" event={"ID":"759333d0-7fcf-4c0b-8733-083e9d5fcc09","Type":"ContainerDied","Data":"f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7"} Nov 24 09:59:46 crc kubenswrapper[4944]: I1124 09:59:46.123780 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx8k4" event={"ID":"759333d0-7fcf-4c0b-8733-083e9d5fcc09","Type":"ContainerStarted","Data":"c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34"} Nov 24 09:59:46 crc kubenswrapper[4944]: I1124 09:59:46.145216 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lx8k4" podStartSLOduration=2.70571672 podStartE2EDuration="5.145194024s" podCreationTimestamp="2025-11-24 09:59:41 +0000 UTC" firstStartedPulling="2025-11-24 09:59:43.079658595 +0000 UTC m=+4043.614099057" lastFinishedPulling="2025-11-24 09:59:45.519135899 +0000 UTC m=+4046.053576361" observedRunningTime="2025-11-24 09:59:46.139067659 +0000 UTC m=+4046.673508131" watchObservedRunningTime="2025-11-24 09:59:46.145194024 +0000 UTC m=+4046.679634486" Nov 24 09:59:51 crc kubenswrapper[4944]: I1124 09:59:51.421787 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:51 crc kubenswrapper[4944]: I1124 09:59:51.422290 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:51 crc kubenswrapper[4944]: I1124 09:59:51.460920 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:52 crc kubenswrapper[4944]: I1124 09:59:52.199637 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:52 crc kubenswrapper[4944]: I1124 09:59:52.236651 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lx8k4"] Nov 24 09:59:53 crc kubenswrapper[4944]: I1124 09:59:53.549025 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:59:53 crc kubenswrapper[4944]: I1124 09:59:53.549114 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:59:54 crc kubenswrapper[4944]: I1124 09:59:54.170934 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lx8k4" podUID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" containerName="registry-server" containerID="cri-o://c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34" gracePeriod=2 Nov 24 09:59:54 crc kubenswrapper[4944]: I1124 09:59:54.555684 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:54 crc kubenswrapper[4944]: I1124 09:59:54.699997 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-catalog-content\") pod \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " Nov 24 09:59:54 crc kubenswrapper[4944]: I1124 09:59:54.700137 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9nv5\" (UniqueName: \"kubernetes.io/projected/759333d0-7fcf-4c0b-8733-083e9d5fcc09-kube-api-access-w9nv5\") pod \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " Nov 24 09:59:54 crc kubenswrapper[4944]: I1124 09:59:54.700193 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-utilities\") pod \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\" (UID: \"759333d0-7fcf-4c0b-8733-083e9d5fcc09\") " Nov 24 09:59:54 crc kubenswrapper[4944]: I1124 09:59:54.701648 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-utilities" (OuterVolumeSpecName: "utilities") pod "759333d0-7fcf-4c0b-8733-083e9d5fcc09" (UID: "759333d0-7fcf-4c0b-8733-083e9d5fcc09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:59:54 crc kubenswrapper[4944]: I1124 09:59:54.701881 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:59:54 crc kubenswrapper[4944]: I1124 09:59:54.708941 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/759333d0-7fcf-4c0b-8733-083e9d5fcc09-kube-api-access-w9nv5" (OuterVolumeSpecName: "kube-api-access-w9nv5") pod "759333d0-7fcf-4c0b-8733-083e9d5fcc09" (UID: "759333d0-7fcf-4c0b-8733-083e9d5fcc09"). InnerVolumeSpecName "kube-api-access-w9nv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:59:54 crc kubenswrapper[4944]: I1124 09:59:54.756737 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "759333d0-7fcf-4c0b-8733-083e9d5fcc09" (UID: "759333d0-7fcf-4c0b-8733-083e9d5fcc09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:59:54 crc kubenswrapper[4944]: I1124 09:59:54.802958 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9nv5\" (UniqueName: \"kubernetes.io/projected/759333d0-7fcf-4c0b-8733-083e9d5fcc09-kube-api-access-w9nv5\") on node \"crc\" DevicePath \"\"" Nov 24 09:59:54 crc kubenswrapper[4944]: I1124 09:59:54.803002 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/759333d0-7fcf-4c0b-8733-083e9d5fcc09-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.182994 4944 generic.go:334] "Generic (PLEG): container finished" podID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" containerID="c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34" exitCode=0 Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.183117 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx8k4" event={"ID":"759333d0-7fcf-4c0b-8733-083e9d5fcc09","Type":"ContainerDied","Data":"c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34"} Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.183164 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx8k4" Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.183204 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx8k4" event={"ID":"759333d0-7fcf-4c0b-8733-083e9d5fcc09","Type":"ContainerDied","Data":"a68d581240bab92a77d5dd620e3124aa56debeb655d52aa1a02a31449427cc89"} Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.183236 4944 scope.go:117] "RemoveContainer" containerID="c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34" Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.219695 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lx8k4"] Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.222612 4944 scope.go:117] "RemoveContainer" containerID="f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7" Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.226448 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lx8k4"] Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.250683 4944 scope.go:117] "RemoveContainer" containerID="db5e214823c58c0ff53d7bd41b8bef496d91481f5eb15a053b0ff99134d51a80" Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.270588 4944 scope.go:117] "RemoveContainer" containerID="c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34" Nov 24 09:59:55 crc kubenswrapper[4944]: E1124 09:59:55.271118 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34\": container with ID starting with c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34 not found: ID does not exist" containerID="c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34" Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.271187 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34"} err="failed to get container status \"c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34\": rpc error: code = NotFound desc = could not find container \"c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34\": container with ID starting with c9ba822a93dce2f345ad9fcd915e35a13b00a61a68885f571f08d91354318b34 not found: ID does not exist" Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.271230 4944 scope.go:117] "RemoveContainer" containerID="f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7" Nov 24 09:59:55 crc kubenswrapper[4944]: E1124 09:59:55.271746 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7\": container with ID starting with f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7 not found: ID does not exist" containerID="f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7" Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.271797 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7"} err="failed to get container status \"f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7\": rpc error: code = NotFound desc = could not find container \"f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7\": container with ID starting with f92726c60c46bbe361f321867692f29bc381c92cdf4278e858e2658ea03803c7 not found: ID does not exist" Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.271822 4944 scope.go:117] "RemoveContainer" containerID="db5e214823c58c0ff53d7bd41b8bef496d91481f5eb15a053b0ff99134d51a80" Nov 24 09:59:55 crc kubenswrapper[4944]: E1124 09:59:55.272703 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db5e214823c58c0ff53d7bd41b8bef496d91481f5eb15a053b0ff99134d51a80\": container with ID starting with db5e214823c58c0ff53d7bd41b8bef496d91481f5eb15a053b0ff99134d51a80 not found: ID does not exist" containerID="db5e214823c58c0ff53d7bd41b8bef496d91481f5eb15a053b0ff99134d51a80" Nov 24 09:59:55 crc kubenswrapper[4944]: I1124 09:59:55.272744 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db5e214823c58c0ff53d7bd41b8bef496d91481f5eb15a053b0ff99134d51a80"} err="failed to get container status \"db5e214823c58c0ff53d7bd41b8bef496d91481f5eb15a053b0ff99134d51a80\": rpc error: code = NotFound desc = could not find container \"db5e214823c58c0ff53d7bd41b8bef496d91481f5eb15a053b0ff99134d51a80\": container with ID starting with db5e214823c58c0ff53d7bd41b8bef496d91481f5eb15a053b0ff99134d51a80 not found: ID does not exist" Nov 24 09:59:56 crc kubenswrapper[4944]: I1124 09:59:56.294132 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" path="/var/lib/kubelet/pods/759333d0-7fcf-4c0b-8733-083e9d5fcc09/volumes" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.147207 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4"] Nov 24 10:00:00 crc kubenswrapper[4944]: E1124 10:00:00.148115 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" containerName="extract-content" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.148131 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" containerName="extract-content" Nov 24 10:00:00 crc kubenswrapper[4944]: E1124 10:00:00.148159 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" containerName="extract-utilities" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.148167 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" containerName="extract-utilities" Nov 24 10:00:00 crc kubenswrapper[4944]: E1124 10:00:00.148193 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" containerName="registry-server" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.148201 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" containerName="registry-server" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.148371 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="759333d0-7fcf-4c0b-8733-083e9d5fcc09" containerName="registry-server" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.148973 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.158743 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.158836 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.161227 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4"] Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.282071 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-secret-volume\") pod \"collect-profiles-29399640-kxqg4\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.282125 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-config-volume\") pod \"collect-profiles-29399640-kxqg4\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.282261 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh8tf\" (UniqueName: \"kubernetes.io/projected/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-kube-api-access-sh8tf\") pod \"collect-profiles-29399640-kxqg4\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.383102 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-secret-volume\") pod \"collect-profiles-29399640-kxqg4\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.383160 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-config-volume\") pod \"collect-profiles-29399640-kxqg4\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.383308 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh8tf\" (UniqueName: \"kubernetes.io/projected/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-kube-api-access-sh8tf\") pod \"collect-profiles-29399640-kxqg4\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.384136 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-config-volume\") pod \"collect-profiles-29399640-kxqg4\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.389513 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-secret-volume\") pod \"collect-profiles-29399640-kxqg4\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.398295 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh8tf\" (UniqueName: \"kubernetes.io/projected/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-kube-api-access-sh8tf\") pod \"collect-profiles-29399640-kxqg4\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.470556 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:00 crc kubenswrapper[4944]: I1124 10:00:00.861361 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4"] Nov 24 10:00:01 crc kubenswrapper[4944]: I1124 10:00:01.245644 4944 generic.go:334] "Generic (PLEG): container finished" podID="3afa46ba-4844-49ee-90c1-fc2f85ce40d3" containerID="e501dd114bb8d78c9c67c389a00602d5f3d274931da0360899d90d8c8febe40b" exitCode=0 Nov 24 10:00:01 crc kubenswrapper[4944]: I1124 10:00:01.245717 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" event={"ID":"3afa46ba-4844-49ee-90c1-fc2f85ce40d3","Type":"ContainerDied","Data":"e501dd114bb8d78c9c67c389a00602d5f3d274931da0360899d90d8c8febe40b"} Nov 24 10:00:01 crc kubenswrapper[4944]: I1124 10:00:01.246028 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" event={"ID":"3afa46ba-4844-49ee-90c1-fc2f85ce40d3","Type":"ContainerStarted","Data":"e6f3401ea37087759414fe754f24e35036c82b80eac29a362e752b791362e595"} Nov 24 10:00:02 crc kubenswrapper[4944]: I1124 10:00:02.492018 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:02 crc kubenswrapper[4944]: I1124 10:00:02.615701 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh8tf\" (UniqueName: \"kubernetes.io/projected/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-kube-api-access-sh8tf\") pod \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " Nov 24 10:00:02 crc kubenswrapper[4944]: I1124 10:00:02.615772 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-config-volume\") pod \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " Nov 24 10:00:02 crc kubenswrapper[4944]: I1124 10:00:02.615817 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-secret-volume\") pod \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\" (UID: \"3afa46ba-4844-49ee-90c1-fc2f85ce40d3\") " Nov 24 10:00:02 crc kubenswrapper[4944]: I1124 10:00:02.616689 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-config-volume" (OuterVolumeSpecName: "config-volume") pod "3afa46ba-4844-49ee-90c1-fc2f85ce40d3" (UID: "3afa46ba-4844-49ee-90c1-fc2f85ce40d3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:00:02 crc kubenswrapper[4944]: I1124 10:00:02.624235 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3afa46ba-4844-49ee-90c1-fc2f85ce40d3" (UID: "3afa46ba-4844-49ee-90c1-fc2f85ce40d3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:00:02 crc kubenswrapper[4944]: I1124 10:00:02.625614 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-kube-api-access-sh8tf" (OuterVolumeSpecName: "kube-api-access-sh8tf") pod "3afa46ba-4844-49ee-90c1-fc2f85ce40d3" (UID: "3afa46ba-4844-49ee-90c1-fc2f85ce40d3"). InnerVolumeSpecName "kube-api-access-sh8tf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:00:02 crc kubenswrapper[4944]: I1124 10:00:02.717767 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh8tf\" (UniqueName: \"kubernetes.io/projected/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-kube-api-access-sh8tf\") on node \"crc\" DevicePath \"\"" Nov 24 10:00:02 crc kubenswrapper[4944]: I1124 10:00:02.717813 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 10:00:02 crc kubenswrapper[4944]: I1124 10:00:02.717825 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3afa46ba-4844-49ee-90c1-fc2f85ce40d3-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 10:00:03 crc kubenswrapper[4944]: I1124 10:00:03.265683 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" event={"ID":"3afa46ba-4844-49ee-90c1-fc2f85ce40d3","Type":"ContainerDied","Data":"e6f3401ea37087759414fe754f24e35036c82b80eac29a362e752b791362e595"} Nov 24 10:00:03 crc kubenswrapper[4944]: I1124 10:00:03.265734 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6f3401ea37087759414fe754f24e35036c82b80eac29a362e752b791362e595" Nov 24 10:00:03 crc kubenswrapper[4944]: I1124 10:00:03.265802 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4" Nov 24 10:00:03 crc kubenswrapper[4944]: I1124 10:00:03.577073 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr"] Nov 24 10:00:03 crc kubenswrapper[4944]: I1124 10:00:03.581354 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399595-gqrtr"] Nov 24 10:00:04 crc kubenswrapper[4944]: I1124 10:00:04.284586 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2319a23-bfcf-45bc-889d-3c2f1bc8c582" path="/var/lib/kubelet/pods/b2319a23-bfcf-45bc-889d-3c2f1bc8c582/volumes" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.085854 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5s4"] Nov 24 10:00:15 crc kubenswrapper[4944]: E1124 10:00:15.087944 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3afa46ba-4844-49ee-90c1-fc2f85ce40d3" containerName="collect-profiles" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.088020 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3afa46ba-4844-49ee-90c1-fc2f85ce40d3" containerName="collect-profiles" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.088257 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3afa46ba-4844-49ee-90c1-fc2f85ce40d3" containerName="collect-profiles" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.089388 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.098858 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5s4"] Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.135483 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-utilities\") pod \"redhat-marketplace-qs5s4\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.135620 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-catalog-content\") pod \"redhat-marketplace-qs5s4\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.135679 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mpnt\" (UniqueName: \"kubernetes.io/projected/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-kube-api-access-5mpnt\") pod \"redhat-marketplace-qs5s4\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.236861 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-catalog-content\") pod \"redhat-marketplace-qs5s4\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.236984 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mpnt\" (UniqueName: \"kubernetes.io/projected/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-kube-api-access-5mpnt\") pod \"redhat-marketplace-qs5s4\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.237031 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-utilities\") pod \"redhat-marketplace-qs5s4\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.237374 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-catalog-content\") pod \"redhat-marketplace-qs5s4\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.237550 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-utilities\") pod \"redhat-marketplace-qs5s4\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.261676 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mpnt\" (UniqueName: \"kubernetes.io/projected/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-kube-api-access-5mpnt\") pod \"redhat-marketplace-qs5s4\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.428841 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:15 crc kubenswrapper[4944]: I1124 10:00:15.670385 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5s4"] Nov 24 10:00:16 crc kubenswrapper[4944]: I1124 10:00:16.352916 4944 generic.go:334] "Generic (PLEG): container finished" podID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" containerID="40847be05def3644cf436b8d05b8b372d5170514c2f84a0be383f1e4a87e8e0b" exitCode=0 Nov 24 10:00:16 crc kubenswrapper[4944]: I1124 10:00:16.353109 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5s4" event={"ID":"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf","Type":"ContainerDied","Data":"40847be05def3644cf436b8d05b8b372d5170514c2f84a0be383f1e4a87e8e0b"} Nov 24 10:00:16 crc kubenswrapper[4944]: I1124 10:00:16.353232 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5s4" event={"ID":"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf","Type":"ContainerStarted","Data":"16738d31ec054537f1dc7f65a66d31b946375826ae5c2106ad1adb48201c492f"} Nov 24 10:00:17 crc kubenswrapper[4944]: I1124 10:00:17.362655 4944 generic.go:334] "Generic (PLEG): container finished" podID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" containerID="2cc6ad3cd0a0dc075ead14482cefa73cfe0e0e20e4655ceccd31b0d66874613d" exitCode=0 Nov 24 10:00:17 crc kubenswrapper[4944]: I1124 10:00:17.362873 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5s4" event={"ID":"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf","Type":"ContainerDied","Data":"2cc6ad3cd0a0dc075ead14482cefa73cfe0e0e20e4655ceccd31b0d66874613d"} Nov 24 10:00:18 crc kubenswrapper[4944]: I1124 10:00:18.370856 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5s4" event={"ID":"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf","Type":"ContainerStarted","Data":"681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769"} Nov 24 10:00:18 crc kubenswrapper[4944]: I1124 10:00:18.391513 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qs5s4" podStartSLOduration=1.844842028 podStartE2EDuration="3.391467414s" podCreationTimestamp="2025-11-24 10:00:15 +0000 UTC" firstStartedPulling="2025-11-24 10:00:16.355497124 +0000 UTC m=+4076.889937586" lastFinishedPulling="2025-11-24 10:00:17.90212251 +0000 UTC m=+4078.436562972" observedRunningTime="2025-11-24 10:00:18.387388074 +0000 UTC m=+4078.921828546" watchObservedRunningTime="2025-11-24 10:00:18.391467414 +0000 UTC m=+4078.925907876" Nov 24 10:00:23 crc kubenswrapper[4944]: I1124 10:00:23.548504 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:00:23 crc kubenswrapper[4944]: I1124 10:00:23.549101 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:00:25 crc kubenswrapper[4944]: I1124 10:00:25.429724 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:25 crc kubenswrapper[4944]: I1124 10:00:25.429767 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:25 crc kubenswrapper[4944]: I1124 10:00:25.468396 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:26 crc kubenswrapper[4944]: I1124 10:00:26.475306 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:26 crc kubenswrapper[4944]: I1124 10:00:26.523246 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5s4"] Nov 24 10:00:28 crc kubenswrapper[4944]: I1124 10:00:28.448351 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qs5s4" podUID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" containerName="registry-server" containerID="cri-o://681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769" gracePeriod=2 Nov 24 10:00:28 crc kubenswrapper[4944]: I1124 10:00:28.817129 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:28 crc kubenswrapper[4944]: I1124 10:00:28.939896 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mpnt\" (UniqueName: \"kubernetes.io/projected/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-kube-api-access-5mpnt\") pod \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " Nov 24 10:00:28 crc kubenswrapper[4944]: I1124 10:00:28.940455 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-utilities\") pod \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " Nov 24 10:00:28 crc kubenswrapper[4944]: I1124 10:00:28.940529 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-catalog-content\") pod \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\" (UID: \"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf\") " Nov 24 10:00:28 crc kubenswrapper[4944]: I1124 10:00:28.941515 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-utilities" (OuterVolumeSpecName: "utilities") pod "9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" (UID: "9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:00:28 crc kubenswrapper[4944]: I1124 10:00:28.945319 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-kube-api-access-5mpnt" (OuterVolumeSpecName: "kube-api-access-5mpnt") pod "9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" (UID: "9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf"). InnerVolumeSpecName "kube-api-access-5mpnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:00:28 crc kubenswrapper[4944]: I1124 10:00:28.957397 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" (UID: "9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.042028 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.042155 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.042171 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mpnt\" (UniqueName: \"kubernetes.io/projected/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf-kube-api-access-5mpnt\") on node \"crc\" DevicePath \"\"" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.558643 4944 generic.go:334] "Generic (PLEG): container finished" podID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" containerID="681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769" exitCode=0 Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.558684 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5s4" event={"ID":"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf","Type":"ContainerDied","Data":"681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769"} Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.558709 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5s4" event={"ID":"9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf","Type":"ContainerDied","Data":"16738d31ec054537f1dc7f65a66d31b946375826ae5c2106ad1adb48201c492f"} Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.558726 4944 scope.go:117] "RemoveContainer" containerID="681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.559392 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs5s4" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.576009 4944 scope.go:117] "RemoveContainer" containerID="2cc6ad3cd0a0dc075ead14482cefa73cfe0e0e20e4655ceccd31b0d66874613d" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.642380 4944 scope.go:117] "RemoveContainer" containerID="40847be05def3644cf436b8d05b8b372d5170514c2f84a0be383f1e4a87e8e0b" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.644364 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5s4"] Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.650329 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5s4"] Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.660694 4944 scope.go:117] "RemoveContainer" containerID="681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769" Nov 24 10:00:29 crc kubenswrapper[4944]: E1124 10:00:29.661102 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769\": container with ID starting with 681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769 not found: ID does not exist" containerID="681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.661128 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769"} err="failed to get container status \"681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769\": rpc error: code = NotFound desc = could not find container \"681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769\": container with ID starting with 681158bbc970e11aaac10ef07ecc78fbf1476c376c2edc7cc8cde75883613769 not found: ID does not exist" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.661148 4944 scope.go:117] "RemoveContainer" containerID="2cc6ad3cd0a0dc075ead14482cefa73cfe0e0e20e4655ceccd31b0d66874613d" Nov 24 10:00:29 crc kubenswrapper[4944]: E1124 10:00:29.661321 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cc6ad3cd0a0dc075ead14482cefa73cfe0e0e20e4655ceccd31b0d66874613d\": container with ID starting with 2cc6ad3cd0a0dc075ead14482cefa73cfe0e0e20e4655ceccd31b0d66874613d not found: ID does not exist" containerID="2cc6ad3cd0a0dc075ead14482cefa73cfe0e0e20e4655ceccd31b0d66874613d" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.661339 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cc6ad3cd0a0dc075ead14482cefa73cfe0e0e20e4655ceccd31b0d66874613d"} err="failed to get container status \"2cc6ad3cd0a0dc075ead14482cefa73cfe0e0e20e4655ceccd31b0d66874613d\": rpc error: code = NotFound desc = could not find container \"2cc6ad3cd0a0dc075ead14482cefa73cfe0e0e20e4655ceccd31b0d66874613d\": container with ID starting with 2cc6ad3cd0a0dc075ead14482cefa73cfe0e0e20e4655ceccd31b0d66874613d not found: ID does not exist" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.661353 4944 scope.go:117] "RemoveContainer" containerID="40847be05def3644cf436b8d05b8b372d5170514c2f84a0be383f1e4a87e8e0b" Nov 24 10:00:29 crc kubenswrapper[4944]: E1124 10:00:29.661522 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40847be05def3644cf436b8d05b8b372d5170514c2f84a0be383f1e4a87e8e0b\": container with ID starting with 40847be05def3644cf436b8d05b8b372d5170514c2f84a0be383f1e4a87e8e0b not found: ID does not exist" containerID="40847be05def3644cf436b8d05b8b372d5170514c2f84a0be383f1e4a87e8e0b" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.661536 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40847be05def3644cf436b8d05b8b372d5170514c2f84a0be383f1e4a87e8e0b"} err="failed to get container status \"40847be05def3644cf436b8d05b8b372d5170514c2f84a0be383f1e4a87e8e0b\": rpc error: code = NotFound desc = could not find container \"40847be05def3644cf436b8d05b8b372d5170514c2f84a0be383f1e4a87e8e0b\": container with ID starting with 40847be05def3644cf436b8d05b8b372d5170514c2f84a0be383f1e4a87e8e0b not found: ID does not exist" Nov 24 10:00:29 crc kubenswrapper[4944]: I1124 10:00:29.678887 4944 scope.go:117] "RemoveContainer" containerID="a2861eff1e5debc3898f43ae95d6265c65e00df89ffe968ea6df630b1443c7c0" Nov 24 10:00:30 crc kubenswrapper[4944]: I1124 10:00:30.286989 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" path="/var/lib/kubelet/pods/9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf/volumes" Nov 24 10:00:53 crc kubenswrapper[4944]: I1124 10:00:53.548992 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:00:53 crc kubenswrapper[4944]: I1124 10:00:53.549539 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:00:53 crc kubenswrapper[4944]: I1124 10:00:53.549599 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 10:00:53 crc kubenswrapper[4944]: I1124 10:00:53.550207 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e5cc635a58a05c32265a03514792334cbb1f6a83a754bd0ae32e207ab72d886d"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 10:00:53 crc kubenswrapper[4944]: I1124 10:00:53.550260 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://e5cc635a58a05c32265a03514792334cbb1f6a83a754bd0ae32e207ab72d886d" gracePeriod=600 Nov 24 10:00:53 crc kubenswrapper[4944]: I1124 10:00:53.713730 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="e5cc635a58a05c32265a03514792334cbb1f6a83a754bd0ae32e207ab72d886d" exitCode=0 Nov 24 10:00:53 crc kubenswrapper[4944]: I1124 10:00:53.713770 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"e5cc635a58a05c32265a03514792334cbb1f6a83a754bd0ae32e207ab72d886d"} Nov 24 10:00:53 crc kubenswrapper[4944]: I1124 10:00:53.713800 4944 scope.go:117] "RemoveContainer" containerID="b5ef6637cd7c9a8f5c61e9cdd367760dbcace73cf74a0ed38b7f069f4558a5fe" Nov 24 10:00:54 crc kubenswrapper[4944]: I1124 10:00:54.724726 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641"} Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.153888 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pmffx"] Nov 24 10:02:13 crc kubenswrapper[4944]: E1124 10:02:13.154883 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" containerName="registry-server" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.154913 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" containerName="registry-server" Nov 24 10:02:13 crc kubenswrapper[4944]: E1124 10:02:13.154926 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" containerName="extract-utilities" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.154933 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" containerName="extract-utilities" Nov 24 10:02:13 crc kubenswrapper[4944]: E1124 10:02:13.154962 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" containerName="extract-content" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.154970 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" containerName="extract-content" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.155159 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ad1da24-84e1-48a1-9a3f-a1c182e8c0cf" containerName="registry-server" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.156517 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.163491 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pmffx"] Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.207146 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-utilities\") pod \"community-operators-pmffx\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.207216 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grpbp\" (UniqueName: \"kubernetes.io/projected/a4303722-96bb-4a35-a2c7-7bad94082d7a-kube-api-access-grpbp\") pod \"community-operators-pmffx\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.207313 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-catalog-content\") pod \"community-operators-pmffx\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.308888 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-catalog-content\") pod \"community-operators-pmffx\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.308996 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-utilities\") pod \"community-operators-pmffx\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.309134 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grpbp\" (UniqueName: \"kubernetes.io/projected/a4303722-96bb-4a35-a2c7-7bad94082d7a-kube-api-access-grpbp\") pod \"community-operators-pmffx\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.311297 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-utilities\") pod \"community-operators-pmffx\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.311125 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-catalog-content\") pod \"community-operators-pmffx\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.334040 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grpbp\" (UniqueName: \"kubernetes.io/projected/a4303722-96bb-4a35-a2c7-7bad94082d7a-kube-api-access-grpbp\") pod \"community-operators-pmffx\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:13 crc kubenswrapper[4944]: I1124 10:02:13.474913 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:14 crc kubenswrapper[4944]: I1124 10:02:14.009371 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pmffx"] Nov 24 10:02:14 crc kubenswrapper[4944]: I1124 10:02:14.451335 4944 generic.go:334] "Generic (PLEG): container finished" podID="a4303722-96bb-4a35-a2c7-7bad94082d7a" containerID="10940863822d73e4717cb202e575fd52903f7481a93d00b1e834f3cbcf80f8a2" exitCode=0 Nov 24 10:02:14 crc kubenswrapper[4944]: I1124 10:02:14.451444 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmffx" event={"ID":"a4303722-96bb-4a35-a2c7-7bad94082d7a","Type":"ContainerDied","Data":"10940863822d73e4717cb202e575fd52903f7481a93d00b1e834f3cbcf80f8a2"} Nov 24 10:02:14 crc kubenswrapper[4944]: I1124 10:02:14.451477 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmffx" event={"ID":"a4303722-96bb-4a35-a2c7-7bad94082d7a","Type":"ContainerStarted","Data":"336f3888e967d0d1948603caf8a97cae3429099db352a0ad72e8e9b3d6ad303a"} Nov 24 10:02:16 crc kubenswrapper[4944]: I1124 10:02:16.480683 4944 generic.go:334] "Generic (PLEG): container finished" podID="a4303722-96bb-4a35-a2c7-7bad94082d7a" containerID="4f6809bd96ebb660452f542d39e12b028edbfccdd9c186a2eb53ec92d042d8da" exitCode=0 Nov 24 10:02:16 crc kubenswrapper[4944]: I1124 10:02:16.480757 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmffx" event={"ID":"a4303722-96bb-4a35-a2c7-7bad94082d7a","Type":"ContainerDied","Data":"4f6809bd96ebb660452f542d39e12b028edbfccdd9c186a2eb53ec92d042d8da"} Nov 24 10:02:17 crc kubenswrapper[4944]: I1124 10:02:17.490309 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmffx" event={"ID":"a4303722-96bb-4a35-a2c7-7bad94082d7a","Type":"ContainerStarted","Data":"9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d"} Nov 24 10:02:17 crc kubenswrapper[4944]: I1124 10:02:17.506084 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pmffx" podStartSLOduration=2.043580441 podStartE2EDuration="4.506066966s" podCreationTimestamp="2025-11-24 10:02:13 +0000 UTC" firstStartedPulling="2025-11-24 10:02:14.453846832 +0000 UTC m=+4194.988287294" lastFinishedPulling="2025-11-24 10:02:16.916333357 +0000 UTC m=+4197.450773819" observedRunningTime="2025-11-24 10:02:17.504567999 +0000 UTC m=+4198.039008471" watchObservedRunningTime="2025-11-24 10:02:17.506066966 +0000 UTC m=+4198.040507428" Nov 24 10:02:23 crc kubenswrapper[4944]: I1124 10:02:23.475154 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:23 crc kubenswrapper[4944]: I1124 10:02:23.475705 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:23 crc kubenswrapper[4944]: I1124 10:02:23.531881 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:23 crc kubenswrapper[4944]: I1124 10:02:23.578338 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:23 crc kubenswrapper[4944]: I1124 10:02:23.760949 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pmffx"] Nov 24 10:02:25 crc kubenswrapper[4944]: I1124 10:02:25.543668 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pmffx" podUID="a4303722-96bb-4a35-a2c7-7bad94082d7a" containerName="registry-server" containerID="cri-o://9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d" gracePeriod=2 Nov 24 10:02:25 crc kubenswrapper[4944]: I1124 10:02:25.912925 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.090874 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-catalog-content\") pod \"a4303722-96bb-4a35-a2c7-7bad94082d7a\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.090936 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-utilities\") pod \"a4303722-96bb-4a35-a2c7-7bad94082d7a\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.091072 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grpbp\" (UniqueName: \"kubernetes.io/projected/a4303722-96bb-4a35-a2c7-7bad94082d7a-kube-api-access-grpbp\") pod \"a4303722-96bb-4a35-a2c7-7bad94082d7a\" (UID: \"a4303722-96bb-4a35-a2c7-7bad94082d7a\") " Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.092518 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-utilities" (OuterVolumeSpecName: "utilities") pod "a4303722-96bb-4a35-a2c7-7bad94082d7a" (UID: "a4303722-96bb-4a35-a2c7-7bad94082d7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.097146 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4303722-96bb-4a35-a2c7-7bad94082d7a-kube-api-access-grpbp" (OuterVolumeSpecName: "kube-api-access-grpbp") pod "a4303722-96bb-4a35-a2c7-7bad94082d7a" (UID: "a4303722-96bb-4a35-a2c7-7bad94082d7a"). InnerVolumeSpecName "kube-api-access-grpbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.148697 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4303722-96bb-4a35-a2c7-7bad94082d7a" (UID: "a4303722-96bb-4a35-a2c7-7bad94082d7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.192930 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.192966 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4303722-96bb-4a35-a2c7-7bad94082d7a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.192982 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grpbp\" (UniqueName: \"kubernetes.io/projected/a4303722-96bb-4a35-a2c7-7bad94082d7a-kube-api-access-grpbp\") on node \"crc\" DevicePath \"\"" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.552350 4944 generic.go:334] "Generic (PLEG): container finished" podID="a4303722-96bb-4a35-a2c7-7bad94082d7a" containerID="9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d" exitCode=0 Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.552407 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmffx" event={"ID":"a4303722-96bb-4a35-a2c7-7bad94082d7a","Type":"ContainerDied","Data":"9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d"} Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.552442 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pmffx" event={"ID":"a4303722-96bb-4a35-a2c7-7bad94082d7a","Type":"ContainerDied","Data":"336f3888e967d0d1948603caf8a97cae3429099db352a0ad72e8e9b3d6ad303a"} Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.552458 4944 scope.go:117] "RemoveContainer" containerID="9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.552675 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pmffx" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.576390 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pmffx"] Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.578682 4944 scope.go:117] "RemoveContainer" containerID="4f6809bd96ebb660452f542d39e12b028edbfccdd9c186a2eb53ec92d042d8da" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.581590 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pmffx"] Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.595105 4944 scope.go:117] "RemoveContainer" containerID="10940863822d73e4717cb202e575fd52903f7481a93d00b1e834f3cbcf80f8a2" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.628624 4944 scope.go:117] "RemoveContainer" containerID="9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d" Nov 24 10:02:26 crc kubenswrapper[4944]: E1124 10:02:26.629187 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d\": container with ID starting with 9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d not found: ID does not exist" containerID="9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.629310 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d"} err="failed to get container status \"9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d\": rpc error: code = NotFound desc = could not find container \"9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d\": container with ID starting with 9f8449fa87674567f971bb48b5eba9c2259c67b8c488a8c2adcd7c18a2bb841d not found: ID does not exist" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.629336 4944 scope.go:117] "RemoveContainer" containerID="4f6809bd96ebb660452f542d39e12b028edbfccdd9c186a2eb53ec92d042d8da" Nov 24 10:02:26 crc kubenswrapper[4944]: E1124 10:02:26.629793 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f6809bd96ebb660452f542d39e12b028edbfccdd9c186a2eb53ec92d042d8da\": container with ID starting with 4f6809bd96ebb660452f542d39e12b028edbfccdd9c186a2eb53ec92d042d8da not found: ID does not exist" containerID="4f6809bd96ebb660452f542d39e12b028edbfccdd9c186a2eb53ec92d042d8da" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.629814 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f6809bd96ebb660452f542d39e12b028edbfccdd9c186a2eb53ec92d042d8da"} err="failed to get container status \"4f6809bd96ebb660452f542d39e12b028edbfccdd9c186a2eb53ec92d042d8da\": rpc error: code = NotFound desc = could not find container \"4f6809bd96ebb660452f542d39e12b028edbfccdd9c186a2eb53ec92d042d8da\": container with ID starting with 4f6809bd96ebb660452f542d39e12b028edbfccdd9c186a2eb53ec92d042d8da not found: ID does not exist" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.629828 4944 scope.go:117] "RemoveContainer" containerID="10940863822d73e4717cb202e575fd52903f7481a93d00b1e834f3cbcf80f8a2" Nov 24 10:02:26 crc kubenswrapper[4944]: E1124 10:02:26.630108 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10940863822d73e4717cb202e575fd52903f7481a93d00b1e834f3cbcf80f8a2\": container with ID starting with 10940863822d73e4717cb202e575fd52903f7481a93d00b1e834f3cbcf80f8a2 not found: ID does not exist" containerID="10940863822d73e4717cb202e575fd52903f7481a93d00b1e834f3cbcf80f8a2" Nov 24 10:02:26 crc kubenswrapper[4944]: I1124 10:02:26.630132 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10940863822d73e4717cb202e575fd52903f7481a93d00b1e834f3cbcf80f8a2"} err="failed to get container status \"10940863822d73e4717cb202e575fd52903f7481a93d00b1e834f3cbcf80f8a2\": rpc error: code = NotFound desc = could not find container \"10940863822d73e4717cb202e575fd52903f7481a93d00b1e834f3cbcf80f8a2\": container with ID starting with 10940863822d73e4717cb202e575fd52903f7481a93d00b1e834f3cbcf80f8a2 not found: ID does not exist" Nov 24 10:02:28 crc kubenswrapper[4944]: I1124 10:02:28.284970 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4303722-96bb-4a35-a2c7-7bad94082d7a" path="/var/lib/kubelet/pods/a4303722-96bb-4a35-a2c7-7bad94082d7a/volumes" Nov 24 10:02:53 crc kubenswrapper[4944]: I1124 10:02:53.548500 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:02:53 crc kubenswrapper[4944]: I1124 10:02:53.549848 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:03:23 crc kubenswrapper[4944]: I1124 10:03:23.549029 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:03:23 crc kubenswrapper[4944]: I1124 10:03:23.549634 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:03:53 crc kubenswrapper[4944]: I1124 10:03:53.549087 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:03:53 crc kubenswrapper[4944]: I1124 10:03:53.549804 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:03:53 crc kubenswrapper[4944]: I1124 10:03:53.549876 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 10:03:53 crc kubenswrapper[4944]: I1124 10:03:53.550650 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 10:03:53 crc kubenswrapper[4944]: I1124 10:03:53.550721 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" gracePeriod=600 Nov 24 10:03:53 crc kubenswrapper[4944]: E1124 10:03:53.679858 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:03:54 crc kubenswrapper[4944]: I1124 10:03:54.192603 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" exitCode=0 Nov 24 10:03:54 crc kubenswrapper[4944]: I1124 10:03:54.192658 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641"} Nov 24 10:03:54 crc kubenswrapper[4944]: I1124 10:03:54.192964 4944 scope.go:117] "RemoveContainer" containerID="e5cc635a58a05c32265a03514792334cbb1f6a83a754bd0ae32e207ab72d886d" Nov 24 10:03:54 crc kubenswrapper[4944]: I1124 10:03:54.193795 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:03:54 crc kubenswrapper[4944]: E1124 10:03:54.194077 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:04:07 crc kubenswrapper[4944]: I1124 10:04:07.277200 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:04:07 crc kubenswrapper[4944]: E1124 10:04:07.278001 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:04:18 crc kubenswrapper[4944]: I1124 10:04:18.276747 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:04:18 crc kubenswrapper[4944]: E1124 10:04:18.277655 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:04:33 crc kubenswrapper[4944]: I1124 10:04:33.277066 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:04:33 crc kubenswrapper[4944]: E1124 10:04:33.277903 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:04:46 crc kubenswrapper[4944]: I1124 10:04:46.276959 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:04:46 crc kubenswrapper[4944]: E1124 10:04:46.277740 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:04:49 crc kubenswrapper[4944]: I1124 10:04:49.922508 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-462nz"] Nov 24 10:04:49 crc kubenswrapper[4944]: I1124 10:04:49.928586 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-462nz"] Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.031436 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-27wwg"] Nov 24 10:04:50 crc kubenswrapper[4944]: E1124 10:04:50.031712 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4303722-96bb-4a35-a2c7-7bad94082d7a" containerName="extract-utilities" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.031724 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4303722-96bb-4a35-a2c7-7bad94082d7a" containerName="extract-utilities" Nov 24 10:04:50 crc kubenswrapper[4944]: E1124 10:04:50.031737 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4303722-96bb-4a35-a2c7-7bad94082d7a" containerName="registry-server" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.031743 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4303722-96bb-4a35-a2c7-7bad94082d7a" containerName="registry-server" Nov 24 10:04:50 crc kubenswrapper[4944]: E1124 10:04:50.031769 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4303722-96bb-4a35-a2c7-7bad94082d7a" containerName="extract-content" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.031777 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4303722-96bb-4a35-a2c7-7bad94082d7a" containerName="extract-content" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.031898 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4303722-96bb-4a35-a2c7-7bad94082d7a" containerName="registry-server" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.032405 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.034491 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.034890 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.035249 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.039537 4944 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-8fr8r" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.044782 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-27wwg"] Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.111133 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/021c1361-2f3f-49b9-925b-aedc59ac3887-node-mnt\") pod \"crc-storage-crc-27wwg\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.111216 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9jq7\" (UniqueName: \"kubernetes.io/projected/021c1361-2f3f-49b9-925b-aedc59ac3887-kube-api-access-f9jq7\") pod \"crc-storage-crc-27wwg\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.111267 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/021c1361-2f3f-49b9-925b-aedc59ac3887-crc-storage\") pod \"crc-storage-crc-27wwg\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.212939 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/021c1361-2f3f-49b9-925b-aedc59ac3887-node-mnt\") pod \"crc-storage-crc-27wwg\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.213488 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9jq7\" (UniqueName: \"kubernetes.io/projected/021c1361-2f3f-49b9-925b-aedc59ac3887-kube-api-access-f9jq7\") pod \"crc-storage-crc-27wwg\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.213546 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/021c1361-2f3f-49b9-925b-aedc59ac3887-crc-storage\") pod \"crc-storage-crc-27wwg\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.213338 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/021c1361-2f3f-49b9-925b-aedc59ac3887-node-mnt\") pod \"crc-storage-crc-27wwg\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.214256 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/021c1361-2f3f-49b9-925b-aedc59ac3887-crc-storage\") pod \"crc-storage-crc-27wwg\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.234085 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9jq7\" (UniqueName: \"kubernetes.io/projected/021c1361-2f3f-49b9-925b-aedc59ac3887-kube-api-access-f9jq7\") pod \"crc-storage-crc-27wwg\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.286401 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7372c865-b00f-4772-89c8-3e5a8a662eb6" path="/var/lib/kubelet/pods/7372c865-b00f-4772-89c8-3e5a8a662eb6/volumes" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.354419 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.792131 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-27wwg"] Nov 24 10:04:50 crc kubenswrapper[4944]: I1124 10:04:50.797909 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 10:04:51 crc kubenswrapper[4944]: I1124 10:04:51.596215 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-27wwg" event={"ID":"021c1361-2f3f-49b9-925b-aedc59ac3887","Type":"ContainerStarted","Data":"1face7647d6c31173d7d8661432523c4108f050516f757d27a0c690b660b3c8c"} Nov 24 10:04:51 crc kubenswrapper[4944]: I1124 10:04:51.596788 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-27wwg" event={"ID":"021c1361-2f3f-49b9-925b-aedc59ac3887","Type":"ContainerStarted","Data":"d89bde5bd1e36ef0e5724f4e09e0bd18ea3da7cdfeeb191882b6e328a50f123d"} Nov 24 10:04:51 crc kubenswrapper[4944]: I1124 10:04:51.618311 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-27wwg" podStartSLOduration=1.24049529 podStartE2EDuration="1.618287481s" podCreationTimestamp="2025-11-24 10:04:50 +0000 UTC" firstStartedPulling="2025-11-24 10:04:50.797499102 +0000 UTC m=+4351.331939604" lastFinishedPulling="2025-11-24 10:04:51.175291333 +0000 UTC m=+4351.709731795" observedRunningTime="2025-11-24 10:04:51.616747071 +0000 UTC m=+4352.151187553" watchObservedRunningTime="2025-11-24 10:04:51.618287481 +0000 UTC m=+4352.152727963" Nov 24 10:04:52 crc kubenswrapper[4944]: I1124 10:04:52.604306 4944 generic.go:334] "Generic (PLEG): container finished" podID="021c1361-2f3f-49b9-925b-aedc59ac3887" containerID="1face7647d6c31173d7d8661432523c4108f050516f757d27a0c690b660b3c8c" exitCode=0 Nov 24 10:04:52 crc kubenswrapper[4944]: I1124 10:04:52.604356 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-27wwg" event={"ID":"021c1361-2f3f-49b9-925b-aedc59ac3887","Type":"ContainerDied","Data":"1face7647d6c31173d7d8661432523c4108f050516f757d27a0c690b660b3c8c"} Nov 24 10:04:53 crc kubenswrapper[4944]: I1124 10:04:53.928122 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.069004 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/021c1361-2f3f-49b9-925b-aedc59ac3887-node-mnt\") pod \"021c1361-2f3f-49b9-925b-aedc59ac3887\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.069120 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9jq7\" (UniqueName: \"kubernetes.io/projected/021c1361-2f3f-49b9-925b-aedc59ac3887-kube-api-access-f9jq7\") pod \"021c1361-2f3f-49b9-925b-aedc59ac3887\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.069139 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/021c1361-2f3f-49b9-925b-aedc59ac3887-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "021c1361-2f3f-49b9-925b-aedc59ac3887" (UID: "021c1361-2f3f-49b9-925b-aedc59ac3887"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.069209 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/021c1361-2f3f-49b9-925b-aedc59ac3887-crc-storage\") pod \"021c1361-2f3f-49b9-925b-aedc59ac3887\" (UID: \"021c1361-2f3f-49b9-925b-aedc59ac3887\") " Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.069436 4944 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/021c1361-2f3f-49b9-925b-aedc59ac3887-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.075222 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/021c1361-2f3f-49b9-925b-aedc59ac3887-kube-api-access-f9jq7" (OuterVolumeSpecName: "kube-api-access-f9jq7") pod "021c1361-2f3f-49b9-925b-aedc59ac3887" (UID: "021c1361-2f3f-49b9-925b-aedc59ac3887"). InnerVolumeSpecName "kube-api-access-f9jq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.092623 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021c1361-2f3f-49b9-925b-aedc59ac3887-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "021c1361-2f3f-49b9-925b-aedc59ac3887" (UID: "021c1361-2f3f-49b9-925b-aedc59ac3887"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.171088 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9jq7\" (UniqueName: \"kubernetes.io/projected/021c1361-2f3f-49b9-925b-aedc59ac3887-kube-api-access-f9jq7\") on node \"crc\" DevicePath \"\"" Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.171123 4944 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/021c1361-2f3f-49b9-925b-aedc59ac3887-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.620626 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-27wwg" event={"ID":"021c1361-2f3f-49b9-925b-aedc59ac3887","Type":"ContainerDied","Data":"d89bde5bd1e36ef0e5724f4e09e0bd18ea3da7cdfeeb191882b6e328a50f123d"} Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.620674 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d89bde5bd1e36ef0e5724f4e09e0bd18ea3da7cdfeeb191882b6e328a50f123d" Nov 24 10:04:54 crc kubenswrapper[4944]: I1124 10:04:54.620705 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-27wwg" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.661225 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-27wwg"] Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.666117 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-27wwg"] Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.769289 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-8bdd7"] Nov 24 10:04:55 crc kubenswrapper[4944]: E1124 10:04:55.769939 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="021c1361-2f3f-49b9-925b-aedc59ac3887" containerName="storage" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.770037 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="021c1361-2f3f-49b9-925b-aedc59ac3887" containerName="storage" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.770304 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="021c1361-2f3f-49b9-925b-aedc59ac3887" containerName="storage" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.771119 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.773376 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.773536 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.773629 4944 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-8fr8r" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.773536 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.775493 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-8bdd7"] Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.894473 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-crc-storage\") pod \"crc-storage-crc-8bdd7\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.894532 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nsjj\" (UniqueName: \"kubernetes.io/projected/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-kube-api-access-4nsjj\") pod \"crc-storage-crc-8bdd7\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.894607 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-node-mnt\") pod \"crc-storage-crc-8bdd7\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.996297 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-crc-storage\") pod \"crc-storage-crc-8bdd7\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.996372 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nsjj\" (UniqueName: \"kubernetes.io/projected/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-kube-api-access-4nsjj\") pod \"crc-storage-crc-8bdd7\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.996468 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-node-mnt\") pod \"crc-storage-crc-8bdd7\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.996734 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-node-mnt\") pod \"crc-storage-crc-8bdd7\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:55 crc kubenswrapper[4944]: I1124 10:04:55.997101 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-crc-storage\") pod \"crc-storage-crc-8bdd7\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:56 crc kubenswrapper[4944]: I1124 10:04:56.014526 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nsjj\" (UniqueName: \"kubernetes.io/projected/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-kube-api-access-4nsjj\") pod \"crc-storage-crc-8bdd7\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:56 crc kubenswrapper[4944]: I1124 10:04:56.086056 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:56 crc kubenswrapper[4944]: I1124 10:04:56.286952 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="021c1361-2f3f-49b9-925b-aedc59ac3887" path="/var/lib/kubelet/pods/021c1361-2f3f-49b9-925b-aedc59ac3887/volumes" Nov 24 10:04:56 crc kubenswrapper[4944]: I1124 10:04:56.543625 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-8bdd7"] Nov 24 10:04:56 crc kubenswrapper[4944]: I1124 10:04:56.634588 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8bdd7" event={"ID":"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14","Type":"ContainerStarted","Data":"b97151fbc3b012d349cb61a2df27aef262158252c812fb012930d424c261cb68"} Nov 24 10:04:57 crc kubenswrapper[4944]: I1124 10:04:57.647996 4944 generic.go:334] "Generic (PLEG): container finished" podID="cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14" containerID="f930847a91aa3a61f223f374ecf57fcd6f92f8bf11b190f7e8e340902278cd06" exitCode=0 Nov 24 10:04:57 crc kubenswrapper[4944]: I1124 10:04:57.648091 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8bdd7" event={"ID":"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14","Type":"ContainerDied","Data":"f930847a91aa3a61f223f374ecf57fcd6f92f8bf11b190f7e8e340902278cd06"} Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.023410 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.145868 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-crc-storage\") pod \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.145991 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-node-mnt\") pod \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.146177 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nsjj\" (UniqueName: \"kubernetes.io/projected/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-kube-api-access-4nsjj\") pod \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\" (UID: \"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14\") " Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.146181 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14" (UID: "cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.146426 4944 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.151613 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-kube-api-access-4nsjj" (OuterVolumeSpecName: "kube-api-access-4nsjj") pod "cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14" (UID: "cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14"). InnerVolumeSpecName "kube-api-access-4nsjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.168851 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14" (UID: "cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.248138 4944 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.248187 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nsjj\" (UniqueName: \"kubernetes.io/projected/cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14-kube-api-access-4nsjj\") on node \"crc\" DevicePath \"\"" Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.277199 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:04:59 crc kubenswrapper[4944]: E1124 10:04:59.277447 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.665392 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8bdd7" event={"ID":"cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14","Type":"ContainerDied","Data":"b97151fbc3b012d349cb61a2df27aef262158252c812fb012930d424c261cb68"} Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.665452 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8bdd7" Nov 24 10:04:59 crc kubenswrapper[4944]: I1124 10:04:59.665457 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b97151fbc3b012d349cb61a2df27aef262158252c812fb012930d424c261cb68" Nov 24 10:05:11 crc kubenswrapper[4944]: I1124 10:05:11.276642 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:05:11 crc kubenswrapper[4944]: E1124 10:05:11.277334 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:05:23 crc kubenswrapper[4944]: I1124 10:05:23.276796 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:05:23 crc kubenswrapper[4944]: E1124 10:05:23.277556 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:05:29 crc kubenswrapper[4944]: I1124 10:05:29.805993 4944 scope.go:117] "RemoveContainer" containerID="9a15994acf814d96da9fbcb84987a2edeb10af89e1b301670246344f9659e481" Nov 24 10:05:37 crc kubenswrapper[4944]: I1124 10:05:37.276974 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:05:37 crc kubenswrapper[4944]: E1124 10:05:37.277677 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:05:49 crc kubenswrapper[4944]: I1124 10:05:49.276990 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:05:49 crc kubenswrapper[4944]: E1124 10:05:49.277676 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:06:00 crc kubenswrapper[4944]: I1124 10:06:00.281374 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:06:00 crc kubenswrapper[4944]: E1124 10:06:00.282023 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:06:15 crc kubenswrapper[4944]: I1124 10:06:15.277030 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:06:15 crc kubenswrapper[4944]: E1124 10:06:15.277845 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:06:29 crc kubenswrapper[4944]: I1124 10:06:29.276531 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:06:29 crc kubenswrapper[4944]: E1124 10:06:29.277228 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:06:44 crc kubenswrapper[4944]: I1124 10:06:44.277941 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:06:44 crc kubenswrapper[4944]: E1124 10:06:44.278767 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:06:58 crc kubenswrapper[4944]: I1124 10:06:58.276965 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:06:58 crc kubenswrapper[4944]: E1124 10:06:58.278615 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:07:12 crc kubenswrapper[4944]: I1124 10:07:12.276624 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:07:12 crc kubenswrapper[4944]: E1124 10:07:12.278495 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:07:23 crc kubenswrapper[4944]: I1124 10:07:23.277405 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:07:23 crc kubenswrapper[4944]: E1124 10:07:23.277968 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:07:36 crc kubenswrapper[4944]: I1124 10:07:36.276826 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:07:36 crc kubenswrapper[4944]: E1124 10:07:36.277605 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:07:47 crc kubenswrapper[4944]: I1124 10:07:47.276290 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:07:47 crc kubenswrapper[4944]: E1124 10:07:47.277049 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:07:58 crc kubenswrapper[4944]: I1124 10:07:58.277334 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:07:58 crc kubenswrapper[4944]: E1124 10:07:58.278270 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.539263 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-2s9fl"] Nov 24 10:08:06 crc kubenswrapper[4944]: E1124 10:08:06.540944 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14" containerName="storage" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.541068 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14" containerName="storage" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.541369 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb1d3c9d-b6ae-4e6a-97ca-08faf4883c14" containerName="storage" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.542290 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.545291 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.545807 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.545990 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.548029 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-rmjf9" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.548312 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.549387 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-config\") pod \"dnsmasq-dns-5d7b5456f5-2s9fl\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.549444 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpgfc\" (UniqueName: \"kubernetes.io/projected/acf125af-3025-4f79-8f2e-948d7b338004-kube-api-access-qpgfc\") pod \"dnsmasq-dns-5d7b5456f5-2s9fl\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.549482 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-2s9fl\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.549681 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-2s9fl"] Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.650213 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-2s9fl\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.650695 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-config\") pod \"dnsmasq-dns-5d7b5456f5-2s9fl\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.650832 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpgfc\" (UniqueName: \"kubernetes.io/projected/acf125af-3025-4f79-8f2e-948d7b338004-kube-api-access-qpgfc\") pod \"dnsmasq-dns-5d7b5456f5-2s9fl\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.651428 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-2s9fl\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.651650 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-config\") pod \"dnsmasq-dns-5d7b5456f5-2s9fl\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.683155 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpgfc\" (UniqueName: \"kubernetes.io/projected/acf125af-3025-4f79-8f2e-948d7b338004-kube-api-access-qpgfc\") pod \"dnsmasq-dns-5d7b5456f5-2s9fl\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.821673 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-qczm9"] Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.824977 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.841904 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-qczm9"] Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.852915 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-config\") pod \"dnsmasq-dns-98ddfc8f-qczm9\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.853010 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97lwx\" (UniqueName: \"kubernetes.io/projected/6a9b2075-dce3-4d8c-8ef4-07c885521442-kube-api-access-97lwx\") pod \"dnsmasq-dns-98ddfc8f-qczm9\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.853065 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-qczm9\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.868115 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.954287 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97lwx\" (UniqueName: \"kubernetes.io/projected/6a9b2075-dce3-4d8c-8ef4-07c885521442-kube-api-access-97lwx\") pod \"dnsmasq-dns-98ddfc8f-qczm9\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.954344 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-qczm9\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.954413 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-config\") pod \"dnsmasq-dns-98ddfc8f-qczm9\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.955699 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-qczm9\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.955862 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-config\") pod \"dnsmasq-dns-98ddfc8f-qczm9\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:06 crc kubenswrapper[4944]: I1124 10:08:06.983173 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97lwx\" (UniqueName: \"kubernetes.io/projected/6a9b2075-dce3-4d8c-8ef4-07c885521442-kube-api-access-97lwx\") pod \"dnsmasq-dns-98ddfc8f-qczm9\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.148112 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.334754 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-2s9fl"] Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.376862 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-qczm9"] Nov 24 10:08:07 crc kubenswrapper[4944]: W1124 10:08:07.385888 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a9b2075_dce3_4d8c_8ef4_07c885521442.slice/crio-68cd8314b624e13c779aea254d7ef4d2e4275cf47fc45bf575290e2f7bd1a053 WatchSource:0}: Error finding container 68cd8314b624e13c779aea254d7ef4d2e4275cf47fc45bf575290e2f7bd1a053: Status 404 returned error can't find the container with id 68cd8314b624e13c779aea254d7ef4d2e4275cf47fc45bf575290e2f7bd1a053 Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.692303 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.698014 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.706247 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.706439 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-tqqcc" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.706738 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.706927 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.707016 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.709583 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.871015 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.871104 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.871123 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcxsd\" (UniqueName: \"kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-kube-api-access-tcxsd\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.871149 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.871188 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a3e1b36d-eda4-4ac1-a417-afada84f0505-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.871211 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.871242 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.871260 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a3e1b36d-eda4-4ac1-a417-afada84f0505-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.871285 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.951446 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.952664 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.955719 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.956408 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.956465 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.968371 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.968389 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-q66nf" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.975357 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.975401 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a3e1b36d-eda4-4ac1-a417-afada84f0505-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.975432 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.975457 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.975498 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.975518 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcxsd\" (UniqueName: \"kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-kube-api-access-tcxsd\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.975541 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.975584 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a3e1b36d-eda4-4ac1-a417-afada84f0505-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.975609 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.976508 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.976800 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.983096 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.984158 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.986365 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 10:08:07 crc kubenswrapper[4944]: I1124 10:08:07.988465 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.000283 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.000330 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/99eac479330c62c6c6b54cc2cb51a2701c71944c1e091963607c33a2179a57cb/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.002202 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a3e1b36d-eda4-4ac1-a417-afada84f0505-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.009215 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcxsd\" (UniqueName: \"kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-kube-api-access-tcxsd\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.015620 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a3e1b36d-eda4-4ac1-a417-afada84f0505-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.029567 4944 generic.go:334] "Generic (PLEG): container finished" podID="6a9b2075-dce3-4d8c-8ef4-07c885521442" containerID="bbc96569ad3797a12427562a1b5f34e4f1b124f2e73c480f7efe294cbe56f625" exitCode=0 Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.029657 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" event={"ID":"6a9b2075-dce3-4d8c-8ef4-07c885521442","Type":"ContainerDied","Data":"bbc96569ad3797a12427562a1b5f34e4f1b124f2e73c480f7efe294cbe56f625"} Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.029690 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" event={"ID":"6a9b2075-dce3-4d8c-8ef4-07c885521442","Type":"ContainerStarted","Data":"68cd8314b624e13c779aea254d7ef4d2e4275cf47fc45bf575290e2f7bd1a053"} Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.031158 4944 generic.go:334] "Generic (PLEG): container finished" podID="acf125af-3025-4f79-8f2e-948d7b338004" containerID="b34a48608906fb8c1f1a0a25e7a4c596701d8a92c4cd358c5459cb87b34b931b" exitCode=0 Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.031191 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" event={"ID":"acf125af-3025-4f79-8f2e-948d7b338004","Type":"ContainerDied","Data":"b34a48608906fb8c1f1a0a25e7a4c596701d8a92c4cd358c5459cb87b34b931b"} Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.031209 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" event={"ID":"acf125af-3025-4f79-8f2e-948d7b338004","Type":"ContainerStarted","Data":"8cdb44c09be18ce9e78db897290378cf36a66da626fa4c6d28b7d8e3ddaf3135"} Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.043857 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\") pod \"rabbitmq-server-0\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " pod="openstack/rabbitmq-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.077240 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.077284 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.077312 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8248999f-3fee-487c-8c29-a07bf54b8954-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.077357 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dvxz\" (UniqueName: \"kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-kube-api-access-2dvxz\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.077373 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8248999f-3fee-487c-8c29-a07bf54b8954-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.077404 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.077432 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.077450 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.077487 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.104771 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.179587 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.179981 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.180027 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8248999f-3fee-487c-8c29-a07bf54b8954-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.180149 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dvxz\" (UniqueName: \"kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-kube-api-access-2dvxz\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.180173 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8248999f-3fee-487c-8c29-a07bf54b8954-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.180232 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.180297 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.180330 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.180420 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.180857 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.181018 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.181712 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.183249 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.185453 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8248999f-3fee-487c-8c29-a07bf54b8954-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.186506 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.187306 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8248999f-3fee-487c-8c29-a07bf54b8954-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.188214 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.188242 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/33b9a91498c17c71a0d96824319b148f89c736699153201088587a2ad180b2b7/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.209554 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dvxz\" (UniqueName: \"kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-kube-api-access-2dvxz\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.255160 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\") pod \"rabbitmq-cell1-server-0\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: E1124 10:08:08.262643 4944 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 24 10:08:08 crc kubenswrapper[4944]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/acf125af-3025-4f79-8f2e-948d7b338004/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 24 10:08:08 crc kubenswrapper[4944]: > podSandboxID="8cdb44c09be18ce9e78db897290378cf36a66da626fa4c6d28b7d8e3ddaf3135" Nov 24 10:08:08 crc kubenswrapper[4944]: E1124 10:08:08.262838 4944 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 24 10:08:08 crc kubenswrapper[4944]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8chc6h5bh56fh546hb7hc8h67h5bchffh577h697h5b5h5bdh59bhf6hf4h558hb5h578h595h5cchfbh644h59ch7fh654h547h587h5cbh5d5h8fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qpgfc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5d7b5456f5-2s9fl_openstack(acf125af-3025-4f79-8f2e-948d7b338004): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/acf125af-3025-4f79-8f2e-948d7b338004/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 24 10:08:08 crc kubenswrapper[4944]: > logger="UnhandledError" Nov 24 10:08:08 crc kubenswrapper[4944]: E1124 10:08:08.267121 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/acf125af-3025-4f79-8f2e-948d7b338004/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" podUID="acf125af-3025-4f79-8f2e-948d7b338004" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.274820 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.520770 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 10:08:08 crc kubenswrapper[4944]: W1124 10:08:08.536532 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8248999f_3fee_487c_8c29_a07bf54b8954.slice/crio-456ce97b15f138e75d856e8fbdbfd13b9e8dd82f7459b21b8463832c43e3c5bc WatchSource:0}: Error finding container 456ce97b15f138e75d856e8fbdbfd13b9e8dd82f7459b21b8463832c43e3c5bc: Status 404 returned error can't find the container with id 456ce97b15f138e75d856e8fbdbfd13b9e8dd82f7459b21b8463832c43e3c5bc Nov 24 10:08:08 crc kubenswrapper[4944]: I1124 10:08:08.620743 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 10:08:08 crc kubenswrapper[4944]: W1124 10:08:08.632967 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3e1b36d_eda4_4ac1_a417_afada84f0505.slice/crio-795ef41bf3c7504dd3edc1231b072a2f39d1d30cab5a4a099b24cdcf0815647d WatchSource:0}: Error finding container 795ef41bf3c7504dd3edc1231b072a2f39d1d30cab5a4a099b24cdcf0815647d: Status 404 returned error can't find the container with id 795ef41bf3c7504dd3edc1231b072a2f39d1d30cab5a4a099b24cdcf0815647d Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.040406 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" event={"ID":"6a9b2075-dce3-4d8c-8ef4-07c885521442","Type":"ContainerStarted","Data":"da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c"} Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.040537 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.041656 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a3e1b36d-eda4-4ac1-a417-afada84f0505","Type":"ContainerStarted","Data":"795ef41bf3c7504dd3edc1231b072a2f39d1d30cab5a4a099b24cdcf0815647d"} Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.042601 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8248999f-3fee-487c-8c29-a07bf54b8954","Type":"ContainerStarted","Data":"456ce97b15f138e75d856e8fbdbfd13b9e8dd82f7459b21b8463832c43e3c5bc"} Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.059324 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" podStartSLOduration=3.059304729 podStartE2EDuration="3.059304729s" podCreationTimestamp="2025-11-24 10:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:08:09.055656143 +0000 UTC m=+4549.590096615" watchObservedRunningTime="2025-11-24 10:08:09.059304729 +0000 UTC m=+4549.593745191" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.211675 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.212889 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.215026 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-qgtrd" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.215427 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.215501 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.215591 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.224526 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.226475 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.297146 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.297883 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.297996 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trhth\" (UniqueName: \"kubernetes.io/projected/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-kube-api-access-trhth\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.298129 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.298204 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-config-data-default\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.298349 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-kolla-config\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.298446 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.299272 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b6593ef6-22dd-4112-996f-a9c575158cee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b6593ef6-22dd-4112-996f-a9c575158cee\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.400634 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-kolla-config\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.400692 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.400748 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b6593ef6-22dd-4112-996f-a9c575158cee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b6593ef6-22dd-4112-996f-a9c575158cee\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.400805 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.400855 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.400885 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trhth\" (UniqueName: \"kubernetes.io/projected/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-kube-api-access-trhth\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.400916 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.400953 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-config-data-default\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.401603 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-kolla-config\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.401640 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.402143 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-config-data-default\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.403237 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.404408 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.404439 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b6593ef6-22dd-4112-996f-a9c575158cee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b6593ef6-22dd-4112-996f-a9c575158cee\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3fec22f162c110e50e167825e930d80613c62452a6502479af13c98a4556520f/globalmount\"" pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.526284 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.527243 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.529557 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-5bgjd" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.529742 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.537598 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.610062 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99n9k\" (UniqueName: \"kubernetes.io/projected/fd7272f9-a6a2-47f1-8d00-f618476ad5b5-kube-api-access-99n9k\") pod \"memcached-0\" (UID: \"fd7272f9-a6a2-47f1-8d00-f618476ad5b5\") " pod="openstack/memcached-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.610109 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fd7272f9-a6a2-47f1-8d00-f618476ad5b5-kolla-config\") pod \"memcached-0\" (UID: \"fd7272f9-a6a2-47f1-8d00-f618476ad5b5\") " pod="openstack/memcached-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.610364 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd7272f9-a6a2-47f1-8d00-f618476ad5b5-config-data\") pod \"memcached-0\" (UID: \"fd7272f9-a6a2-47f1-8d00-f618476ad5b5\") " pod="openstack/memcached-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.696208 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.696287 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.697069 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trhth\" (UniqueName: \"kubernetes.io/projected/1d94b6e3-a5b3-40dd-8346-97a1f5ece983-kube-api-access-trhth\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.711319 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd7272f9-a6a2-47f1-8d00-f618476ad5b5-config-data\") pod \"memcached-0\" (UID: \"fd7272f9-a6a2-47f1-8d00-f618476ad5b5\") " pod="openstack/memcached-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.711452 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99n9k\" (UniqueName: \"kubernetes.io/projected/fd7272f9-a6a2-47f1-8d00-f618476ad5b5-kube-api-access-99n9k\") pod \"memcached-0\" (UID: \"fd7272f9-a6a2-47f1-8d00-f618476ad5b5\") " pod="openstack/memcached-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.711489 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fd7272f9-a6a2-47f1-8d00-f618476ad5b5-kolla-config\") pod \"memcached-0\" (UID: \"fd7272f9-a6a2-47f1-8d00-f618476ad5b5\") " pod="openstack/memcached-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.712326 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fd7272f9-a6a2-47f1-8d00-f618476ad5b5-kolla-config\") pod \"memcached-0\" (UID: \"fd7272f9-a6a2-47f1-8d00-f618476ad5b5\") " pod="openstack/memcached-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.712496 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd7272f9-a6a2-47f1-8d00-f618476ad5b5-config-data\") pod \"memcached-0\" (UID: \"fd7272f9-a6a2-47f1-8d00-f618476ad5b5\") " pod="openstack/memcached-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.732035 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99n9k\" (UniqueName: \"kubernetes.io/projected/fd7272f9-a6a2-47f1-8d00-f618476ad5b5-kube-api-access-99n9k\") pod \"memcached-0\" (UID: \"fd7272f9-a6a2-47f1-8d00-f618476ad5b5\") " pod="openstack/memcached-0" Nov 24 10:08:09 crc kubenswrapper[4944]: I1124 10:08:09.844336 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.106117 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8248999f-3fee-487c-8c29-a07bf54b8954","Type":"ContainerStarted","Data":"90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90"} Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.122969 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" event={"ID":"acf125af-3025-4f79-8f2e-948d7b338004","Type":"ContainerStarted","Data":"9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1"} Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.123232 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.154646 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" podStartSLOduration=4.15462568 podStartE2EDuration="4.15462568s" podCreationTimestamp="2025-11-24 10:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:08:10.150448327 +0000 UTC m=+4550.684888779" watchObservedRunningTime="2025-11-24 10:08:10.15462568 +0000 UTC m=+4550.689066142" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.285112 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:08:10 crc kubenswrapper[4944]: E1124 10:08:10.285648 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.303608 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b6593ef6-22dd-4112-996f-a9c575158cee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b6593ef6-22dd-4112-996f-a9c575158cee\") pod \"openstack-galera-0\" (UID: \"1d94b6e3-a5b3-40dd-8346-97a1f5ece983\") " pod="openstack/openstack-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.408251 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.434343 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.599140 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.602203 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.604124 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.605009 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.605251 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8l6gc" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.606103 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.610025 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.675969 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 10:08:10 crc kubenswrapper[4944]: W1124 10:08:10.679107 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d94b6e3_a5b3_40dd_8346_97a1f5ece983.slice/crio-72856dde9997cd9de0ac886d2b8754eb5729dbf8fcc242fb467377f118bb39bc WatchSource:0}: Error finding container 72856dde9997cd9de0ac886d2b8754eb5729dbf8fcc242fb467377f118bb39bc: Status 404 returned error can't find the container with id 72856dde9997cd9de0ac886d2b8754eb5729dbf8fcc242fb467377f118bb39bc Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.729912 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56ccc568-803f-4862-b9d6-f5e29d45f5b0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.730285 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/56ccc568-803f-4862-b9d6-f5e29d45f5b0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.730307 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snjvv\" (UniqueName: \"kubernetes.io/projected/56ccc568-803f-4862-b9d6-f5e29d45f5b0-kube-api-access-snjvv\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.730396 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/56ccc568-803f-4862-b9d6-f5e29d45f5b0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.730439 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56ccc568-803f-4862-b9d6-f5e29d45f5b0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.730461 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6a26ae50-15c8-4f74-84ce-ce6dbd6d9615\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6a26ae50-15c8-4f74-84ce-ce6dbd6d9615\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.730482 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/56ccc568-803f-4862-b9d6-f5e29d45f5b0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.730507 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56ccc568-803f-4862-b9d6-f5e29d45f5b0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.831878 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/56ccc568-803f-4862-b9d6-f5e29d45f5b0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.831924 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snjvv\" (UniqueName: \"kubernetes.io/projected/56ccc568-803f-4862-b9d6-f5e29d45f5b0-kube-api-access-snjvv\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.831981 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/56ccc568-803f-4862-b9d6-f5e29d45f5b0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.832018 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56ccc568-803f-4862-b9d6-f5e29d45f5b0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.832041 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6a26ae50-15c8-4f74-84ce-ce6dbd6d9615\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6a26ae50-15c8-4f74-84ce-ce6dbd6d9615\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.832156 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/56ccc568-803f-4862-b9d6-f5e29d45f5b0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.832185 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56ccc568-803f-4862-b9d6-f5e29d45f5b0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.832217 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56ccc568-803f-4862-b9d6-f5e29d45f5b0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.833459 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/56ccc568-803f-4862-b9d6-f5e29d45f5b0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.833471 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/56ccc568-803f-4862-b9d6-f5e29d45f5b0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.833896 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56ccc568-803f-4862-b9d6-f5e29d45f5b0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.834240 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/56ccc568-803f-4862-b9d6-f5e29d45f5b0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.835658 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.835697 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6a26ae50-15c8-4f74-84ce-ce6dbd6d9615\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6a26ae50-15c8-4f74-84ce-ce6dbd6d9615\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/852e2b20bc47893945c5419f55595b0de77328a58cb2aa59ed518922753ffa0c/globalmount\"" pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.836366 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/56ccc568-803f-4862-b9d6-f5e29d45f5b0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.837324 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56ccc568-803f-4862-b9d6-f5e29d45f5b0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.849319 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snjvv\" (UniqueName: \"kubernetes.io/projected/56ccc568-803f-4862-b9d6-f5e29d45f5b0-kube-api-access-snjvv\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.858410 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6a26ae50-15c8-4f74-84ce-ce6dbd6d9615\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6a26ae50-15c8-4f74-84ce-ce6dbd6d9615\") pod \"openstack-cell1-galera-0\" (UID: \"56ccc568-803f-4862-b9d6-f5e29d45f5b0\") " pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:10 crc kubenswrapper[4944]: I1124 10:08:10.922640 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:11 crc kubenswrapper[4944]: I1124 10:08:11.129696 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fd7272f9-a6a2-47f1-8d00-f618476ad5b5","Type":"ContainerStarted","Data":"3bd2d48af19b982fb56770fca9a73e23931a3ebbb625925447ec5650aef48632"} Nov 24 10:08:11 crc kubenswrapper[4944]: I1124 10:08:11.130078 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 10:08:11 crc kubenswrapper[4944]: I1124 10:08:11.130095 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fd7272f9-a6a2-47f1-8d00-f618476ad5b5","Type":"ContainerStarted","Data":"5b5a78dd64c7ec16c442e141424d665de43df386a26f17825afa673d56c8ab74"} Nov 24 10:08:11 crc kubenswrapper[4944]: I1124 10:08:11.131556 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d94b6e3-a5b3-40dd-8346-97a1f5ece983","Type":"ContainerStarted","Data":"44a14e819d7571a4da1d3d7a42dd470b89f2402134951b9b68debc78eaf7e01e"} Nov 24 10:08:11 crc kubenswrapper[4944]: I1124 10:08:11.131584 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d94b6e3-a5b3-40dd-8346-97a1f5ece983","Type":"ContainerStarted","Data":"72856dde9997cd9de0ac886d2b8754eb5729dbf8fcc242fb467377f118bb39bc"} Nov 24 10:08:11 crc kubenswrapper[4944]: I1124 10:08:11.132673 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a3e1b36d-eda4-4ac1-a417-afada84f0505","Type":"ContainerStarted","Data":"4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff"} Nov 24 10:08:11 crc kubenswrapper[4944]: I1124 10:08:11.149503 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.149487212 podStartE2EDuration="2.149487212s" podCreationTimestamp="2025-11-24 10:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:08:11.146892989 +0000 UTC m=+4551.681333471" watchObservedRunningTime="2025-11-24 10:08:11.149487212 +0000 UTC m=+4551.683927674" Nov 24 10:08:11 crc kubenswrapper[4944]: I1124 10:08:11.319508 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 10:08:11 crc kubenswrapper[4944]: W1124 10:08:11.399430 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56ccc568_803f_4862_b9d6_f5e29d45f5b0.slice/crio-ab3c9a3848d9bc21f811f2210a99626e81041c21ece3f20f7b484f46bc4fea4b WatchSource:0}: Error finding container ab3c9a3848d9bc21f811f2210a99626e81041c21ece3f20f7b484f46bc4fea4b: Status 404 returned error can't find the container with id ab3c9a3848d9bc21f811f2210a99626e81041c21ece3f20f7b484f46bc4fea4b Nov 24 10:08:12 crc kubenswrapper[4944]: I1124 10:08:12.163531 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"56ccc568-803f-4862-b9d6-f5e29d45f5b0","Type":"ContainerStarted","Data":"f081f7c1de469b9ede1cbfb937fb009fafe1b30d0c8dacc0f5474988adc04320"} Nov 24 10:08:12 crc kubenswrapper[4944]: I1124 10:08:12.163604 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"56ccc568-803f-4862-b9d6-f5e29d45f5b0","Type":"ContainerStarted","Data":"ab3c9a3848d9bc21f811f2210a99626e81041c21ece3f20f7b484f46bc4fea4b"} Nov 24 10:08:15 crc kubenswrapper[4944]: I1124 10:08:15.182963 4944 generic.go:334] "Generic (PLEG): container finished" podID="1d94b6e3-a5b3-40dd-8346-97a1f5ece983" containerID="44a14e819d7571a4da1d3d7a42dd470b89f2402134951b9b68debc78eaf7e01e" exitCode=0 Nov 24 10:08:15 crc kubenswrapper[4944]: I1124 10:08:15.183066 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d94b6e3-a5b3-40dd-8346-97a1f5ece983","Type":"ContainerDied","Data":"44a14e819d7571a4da1d3d7a42dd470b89f2402134951b9b68debc78eaf7e01e"} Nov 24 10:08:16 crc kubenswrapper[4944]: I1124 10:08:16.195108 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1d94b6e3-a5b3-40dd-8346-97a1f5ece983","Type":"ContainerStarted","Data":"2170bf89156101c360d0cd1d540c2fe855c429a8369a5e94e3e10b849a7520f8"} Nov 24 10:08:16 crc kubenswrapper[4944]: I1124 10:08:16.219488 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.219467446 podStartE2EDuration="8.219467446s" podCreationTimestamp="2025-11-24 10:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:08:16.212705351 +0000 UTC m=+4556.747145823" watchObservedRunningTime="2025-11-24 10:08:16.219467446 +0000 UTC m=+4556.753907908" Nov 24 10:08:16 crc kubenswrapper[4944]: I1124 10:08:16.870328 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.149347 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.198471 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-2s9fl"] Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.208076 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"56ccc568-803f-4862-b9d6-f5e29d45f5b0","Type":"ContainerDied","Data":"f081f7c1de469b9ede1cbfb937fb009fafe1b30d0c8dacc0f5474988adc04320"} Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.208004 4944 generic.go:334] "Generic (PLEG): container finished" podID="56ccc568-803f-4862-b9d6-f5e29d45f5b0" containerID="f081f7c1de469b9ede1cbfb937fb009fafe1b30d0c8dacc0f5474988adc04320" exitCode=0 Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.208536 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" podUID="acf125af-3025-4f79-8f2e-948d7b338004" containerName="dnsmasq-dns" containerID="cri-o://9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1" gracePeriod=10 Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.627661 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.756640 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-config\") pod \"acf125af-3025-4f79-8f2e-948d7b338004\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.756746 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpgfc\" (UniqueName: \"kubernetes.io/projected/acf125af-3025-4f79-8f2e-948d7b338004-kube-api-access-qpgfc\") pod \"acf125af-3025-4f79-8f2e-948d7b338004\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.756843 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-dns-svc\") pod \"acf125af-3025-4f79-8f2e-948d7b338004\" (UID: \"acf125af-3025-4f79-8f2e-948d7b338004\") " Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.761750 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acf125af-3025-4f79-8f2e-948d7b338004-kube-api-access-qpgfc" (OuterVolumeSpecName: "kube-api-access-qpgfc") pod "acf125af-3025-4f79-8f2e-948d7b338004" (UID: "acf125af-3025-4f79-8f2e-948d7b338004"). InnerVolumeSpecName "kube-api-access-qpgfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.788039 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-config" (OuterVolumeSpecName: "config") pod "acf125af-3025-4f79-8f2e-948d7b338004" (UID: "acf125af-3025-4f79-8f2e-948d7b338004"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.788628 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "acf125af-3025-4f79-8f2e-948d7b338004" (UID: "acf125af-3025-4f79-8f2e-948d7b338004"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.858568 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.858609 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acf125af-3025-4f79-8f2e-948d7b338004-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:08:17 crc kubenswrapper[4944]: I1124 10:08:17.858626 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpgfc\" (UniqueName: \"kubernetes.io/projected/acf125af-3025-4f79-8f2e-948d7b338004-kube-api-access-qpgfc\") on node \"crc\" DevicePath \"\"" Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.217103 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"56ccc568-803f-4862-b9d6-f5e29d45f5b0","Type":"ContainerStarted","Data":"b794366e4364e387f52da58bfe5cb747ba0767bb29fe701065be97dfb55d83ad"} Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.220017 4944 generic.go:334] "Generic (PLEG): container finished" podID="acf125af-3025-4f79-8f2e-948d7b338004" containerID="9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1" exitCode=0 Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.220080 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" event={"ID":"acf125af-3025-4f79-8f2e-948d7b338004","Type":"ContainerDied","Data":"9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1"} Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.220129 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" event={"ID":"acf125af-3025-4f79-8f2e-948d7b338004","Type":"ContainerDied","Data":"8cdb44c09be18ce9e78db897290378cf36a66da626fa4c6d28b7d8e3ddaf3135"} Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.220165 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-2s9fl" Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.220166 4944 scope.go:117] "RemoveContainer" containerID="9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1" Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.246276 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=9.24625058 podStartE2EDuration="9.24625058s" podCreationTimestamp="2025-11-24 10:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:08:18.2396674 +0000 UTC m=+4558.774107872" watchObservedRunningTime="2025-11-24 10:08:18.24625058 +0000 UTC m=+4558.780691062" Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.267735 4944 scope.go:117] "RemoveContainer" containerID="b34a48608906fb8c1f1a0a25e7a4c596701d8a92c4cd358c5459cb87b34b931b" Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.272561 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-2s9fl"] Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.288352 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-2s9fl"] Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.295449 4944 scope.go:117] "RemoveContainer" containerID="9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1" Nov 24 10:08:18 crc kubenswrapper[4944]: E1124 10:08:18.295875 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1\": container with ID starting with 9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1 not found: ID does not exist" containerID="9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1" Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.295916 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1"} err="failed to get container status \"9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1\": rpc error: code = NotFound desc = could not find container \"9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1\": container with ID starting with 9f90f1fc69ecf72e7a47b2fab8706c94af78ac5cf42850c5454746b1a2962eb1 not found: ID does not exist" Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.295942 4944 scope.go:117] "RemoveContainer" containerID="b34a48608906fb8c1f1a0a25e7a4c596701d8a92c4cd358c5459cb87b34b931b" Nov 24 10:08:18 crc kubenswrapper[4944]: E1124 10:08:18.296215 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b34a48608906fb8c1f1a0a25e7a4c596701d8a92c4cd358c5459cb87b34b931b\": container with ID starting with b34a48608906fb8c1f1a0a25e7a4c596701d8a92c4cd358c5459cb87b34b931b not found: ID does not exist" containerID="b34a48608906fb8c1f1a0a25e7a4c596701d8a92c4cd358c5459cb87b34b931b" Nov 24 10:08:18 crc kubenswrapper[4944]: I1124 10:08:18.296234 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b34a48608906fb8c1f1a0a25e7a4c596701d8a92c4cd358c5459cb87b34b931b"} err="failed to get container status \"b34a48608906fb8c1f1a0a25e7a4c596701d8a92c4cd358c5459cb87b34b931b\": rpc error: code = NotFound desc = could not find container \"b34a48608906fb8c1f1a0a25e7a4c596701d8a92c4cd358c5459cb87b34b931b\": container with ID starting with b34a48608906fb8c1f1a0a25e7a4c596701d8a92c4cd358c5459cb87b34b931b not found: ID does not exist" Nov 24 10:08:19 crc kubenswrapper[4944]: I1124 10:08:19.844943 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 10:08:20 crc kubenswrapper[4944]: I1124 10:08:20.290637 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acf125af-3025-4f79-8f2e-948d7b338004" path="/var/lib/kubelet/pods/acf125af-3025-4f79-8f2e-948d7b338004/volumes" Nov 24 10:08:20 crc kubenswrapper[4944]: I1124 10:08:20.435072 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 10:08:20 crc kubenswrapper[4944]: I1124 10:08:20.435147 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 10:08:20 crc kubenswrapper[4944]: I1124 10:08:20.523164 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 10:08:20 crc kubenswrapper[4944]: E1124 10:08:20.726024 4944 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.129.56.15:33946->38.129.56.15:34895: read tcp 38.129.56.15:33946->38.129.56.15:34895: read: connection reset by peer Nov 24 10:08:20 crc kubenswrapper[4944]: I1124 10:08:20.923670 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:20 crc kubenswrapper[4944]: I1124 10:08:20.923793 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:21 crc kubenswrapper[4944]: I1124 10:08:21.277098 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:08:21 crc kubenswrapper[4944]: E1124 10:08:21.277570 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:08:21 crc kubenswrapper[4944]: I1124 10:08:21.304791 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 10:08:23 crc kubenswrapper[4944]: I1124 10:08:23.120407 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:23 crc kubenswrapper[4944]: I1124 10:08:23.196105 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 10:08:34 crc kubenswrapper[4944]: I1124 10:08:34.277219 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:08:34 crc kubenswrapper[4944]: E1124 10:08:34.278891 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:08:42 crc kubenswrapper[4944]: I1124 10:08:42.410943 4944 generic.go:334] "Generic (PLEG): container finished" podID="8248999f-3fee-487c-8c29-a07bf54b8954" containerID="90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90" exitCode=0 Nov 24 10:08:42 crc kubenswrapper[4944]: I1124 10:08:42.411466 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8248999f-3fee-487c-8c29-a07bf54b8954","Type":"ContainerDied","Data":"90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90"} Nov 24 10:08:43 crc kubenswrapper[4944]: I1124 10:08:43.421462 4944 generic.go:334] "Generic (PLEG): container finished" podID="a3e1b36d-eda4-4ac1-a417-afada84f0505" containerID="4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff" exitCode=0 Nov 24 10:08:43 crc kubenswrapper[4944]: I1124 10:08:43.421551 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a3e1b36d-eda4-4ac1-a417-afada84f0505","Type":"ContainerDied","Data":"4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff"} Nov 24 10:08:43 crc kubenswrapper[4944]: I1124 10:08:43.424216 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8248999f-3fee-487c-8c29-a07bf54b8954","Type":"ContainerStarted","Data":"8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536"} Nov 24 10:08:43 crc kubenswrapper[4944]: I1124 10:08:43.424493 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:08:43 crc kubenswrapper[4944]: I1124 10:08:43.468428 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.46840848 podStartE2EDuration="37.46840848s" podCreationTimestamp="2025-11-24 10:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:08:43.462402428 +0000 UTC m=+4583.996842910" watchObservedRunningTime="2025-11-24 10:08:43.46840848 +0000 UTC m=+4584.002848942" Nov 24 10:08:44 crc kubenswrapper[4944]: I1124 10:08:44.443685 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a3e1b36d-eda4-4ac1-a417-afada84f0505","Type":"ContainerStarted","Data":"da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14"} Nov 24 10:08:44 crc kubenswrapper[4944]: I1124 10:08:44.444739 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 10:08:44 crc kubenswrapper[4944]: I1124 10:08:44.467447 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.467422344 podStartE2EDuration="38.467422344s" podCreationTimestamp="2025-11-24 10:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:08:44.463763707 +0000 UTC m=+4584.998204169" watchObservedRunningTime="2025-11-24 10:08:44.467422344 +0000 UTC m=+4585.001862806" Nov 24 10:08:45 crc kubenswrapper[4944]: I1124 10:08:45.276864 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:08:45 crc kubenswrapper[4944]: E1124 10:08:45.277537 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:08:58 crc kubenswrapper[4944]: I1124 10:08:58.109186 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 10:08:58 crc kubenswrapper[4944]: I1124 10:08:58.286665 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:00 crc kubenswrapper[4944]: I1124 10:09:00.281030 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:09:01 crc kubenswrapper[4944]: I1124 10:09:01.554297 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"54a6b243aeb67222b25b473701ee151808da9cae0e39b5478f62186939a082c4"} Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.620452 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-png4d"] Nov 24 10:09:02 crc kubenswrapper[4944]: E1124 10:09:02.621118 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acf125af-3025-4f79-8f2e-948d7b338004" containerName="init" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.621131 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="acf125af-3025-4f79-8f2e-948d7b338004" containerName="init" Nov 24 10:09:02 crc kubenswrapper[4944]: E1124 10:09:02.621157 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acf125af-3025-4f79-8f2e-948d7b338004" containerName="dnsmasq-dns" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.621163 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="acf125af-3025-4f79-8f2e-948d7b338004" containerName="dnsmasq-dns" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.621325 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="acf125af-3025-4f79-8f2e-948d7b338004" containerName="dnsmasq-dns" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.622184 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.636462 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-png4d"] Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.779276 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2bv4\" (UniqueName: \"kubernetes.io/projected/7ee58980-1615-4769-8d7b-fda64bd2276c-kube-api-access-c2bv4\") pod \"dnsmasq-dns-5b7946d7b9-png4d\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.779392 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-png4d\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.779430 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-config\") pod \"dnsmasq-dns-5b7946d7b9-png4d\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.880872 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2bv4\" (UniqueName: \"kubernetes.io/projected/7ee58980-1615-4769-8d7b-fda64bd2276c-kube-api-access-c2bv4\") pod \"dnsmasq-dns-5b7946d7b9-png4d\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.881288 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-png4d\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.881326 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-config\") pod \"dnsmasq-dns-5b7946d7b9-png4d\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.882063 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-png4d\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.882217 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-config\") pod \"dnsmasq-dns-5b7946d7b9-png4d\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.913829 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2bv4\" (UniqueName: \"kubernetes.io/projected/7ee58980-1615-4769-8d7b-fda64bd2276c-kube-api-access-c2bv4\") pod \"dnsmasq-dns-5b7946d7b9-png4d\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:02 crc kubenswrapper[4944]: I1124 10:09:02.939230 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:03 crc kubenswrapper[4944]: I1124 10:09:03.188085 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 10:09:03 crc kubenswrapper[4944]: I1124 10:09:03.369229 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-png4d"] Nov 24 10:09:03 crc kubenswrapper[4944]: W1124 10:09:03.371615 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ee58980_1615_4769_8d7b_fda64bd2276c.slice/crio-1e51c371b0ca6f8e67680bffd9c155d00cf12a1f5c9fd9738cb686e64bc1077d WatchSource:0}: Error finding container 1e51c371b0ca6f8e67680bffd9c155d00cf12a1f5c9fd9738cb686e64bc1077d: Status 404 returned error can't find the container with id 1e51c371b0ca6f8e67680bffd9c155d00cf12a1f5c9fd9738cb686e64bc1077d Nov 24 10:09:03 crc kubenswrapper[4944]: I1124 10:09:03.567372 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" event={"ID":"7ee58980-1615-4769-8d7b-fda64bd2276c","Type":"ContainerStarted","Data":"636fda1fcd9206ae51ad46629bff26db37ab8d4aa56be5818e3747d3bd3c9486"} Nov 24 10:09:03 crc kubenswrapper[4944]: I1124 10:09:03.567416 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" event={"ID":"7ee58980-1615-4769-8d7b-fda64bd2276c","Type":"ContainerStarted","Data":"1e51c371b0ca6f8e67680bffd9c155d00cf12a1f5c9fd9738cb686e64bc1077d"} Nov 24 10:09:03 crc kubenswrapper[4944]: I1124 10:09:03.748089 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 10:09:04 crc kubenswrapper[4944]: I1124 10:09:04.575661 4944 generic.go:334] "Generic (PLEG): container finished" podID="7ee58980-1615-4769-8d7b-fda64bd2276c" containerID="636fda1fcd9206ae51ad46629bff26db37ab8d4aa56be5818e3747d3bd3c9486" exitCode=0 Nov 24 10:09:04 crc kubenswrapper[4944]: I1124 10:09:04.575709 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" event={"ID":"7ee58980-1615-4769-8d7b-fda64bd2276c","Type":"ContainerDied","Data":"636fda1fcd9206ae51ad46629bff26db37ab8d4aa56be5818e3747d3bd3c9486"} Nov 24 10:09:04 crc kubenswrapper[4944]: I1124 10:09:04.925582 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="a3e1b36d-eda4-4ac1-a417-afada84f0505" containerName="rabbitmq" containerID="cri-o://da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14" gracePeriod=604799 Nov 24 10:09:05 crc kubenswrapper[4944]: I1124 10:09:05.446530 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="8248999f-3fee-487c-8c29-a07bf54b8954" containerName="rabbitmq" containerID="cri-o://8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536" gracePeriod=604799 Nov 24 10:09:05 crc kubenswrapper[4944]: I1124 10:09:05.583907 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" event={"ID":"7ee58980-1615-4769-8d7b-fda64bd2276c","Type":"ContainerStarted","Data":"31c0689fd589e8d8dd1d24c1f5ebd0d951cf9e366b99499a678950a3826e2f6d"} Nov 24 10:09:05 crc kubenswrapper[4944]: I1124 10:09:05.584105 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:05 crc kubenswrapper[4944]: I1124 10:09:05.607094 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" podStartSLOduration=3.607045631 podStartE2EDuration="3.607045631s" podCreationTimestamp="2025-11-24 10:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:09:05.602367431 +0000 UTC m=+4606.136807893" watchObservedRunningTime="2025-11-24 10:09:05.607045631 +0000 UTC m=+4606.141486093" Nov 24 10:09:08 crc kubenswrapper[4944]: I1124 10:09:08.107203 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="a3e1b36d-eda4-4ac1-a417-afada84f0505" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.236:5672: connect: connection refused" Nov 24 10:09:08 crc kubenswrapper[4944]: I1124 10:09:08.277007 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="8248999f-3fee-487c-8c29-a07bf54b8954" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.237:5672: connect: connection refused" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.341704 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.506639 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcxsd\" (UniqueName: \"kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-kube-api-access-tcxsd\") pod \"a3e1b36d-eda4-4ac1-a417-afada84f0505\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.506706 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a3e1b36d-eda4-4ac1-a417-afada84f0505-pod-info\") pod \"a3e1b36d-eda4-4ac1-a417-afada84f0505\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.506746 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a3e1b36d-eda4-4ac1-a417-afada84f0505-erlang-cookie-secret\") pod \"a3e1b36d-eda4-4ac1-a417-afada84f0505\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.506776 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-plugins-conf\") pod \"a3e1b36d-eda4-4ac1-a417-afada84f0505\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.506897 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\") pod \"a3e1b36d-eda4-4ac1-a417-afada84f0505\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.506917 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-server-conf\") pod \"a3e1b36d-eda4-4ac1-a417-afada84f0505\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.506937 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-plugins\") pod \"a3e1b36d-eda4-4ac1-a417-afada84f0505\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.507007 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-erlang-cookie\") pod \"a3e1b36d-eda4-4ac1-a417-afada84f0505\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.507105 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-confd\") pod \"a3e1b36d-eda4-4ac1-a417-afada84f0505\" (UID: \"a3e1b36d-eda4-4ac1-a417-afada84f0505\") " Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.508004 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a3e1b36d-eda4-4ac1-a417-afada84f0505" (UID: "a3e1b36d-eda4-4ac1-a417-afada84f0505"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.508062 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a3e1b36d-eda4-4ac1-a417-afada84f0505" (UID: "a3e1b36d-eda4-4ac1-a417-afada84f0505"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.509100 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a3e1b36d-eda4-4ac1-a417-afada84f0505" (UID: "a3e1b36d-eda4-4ac1-a417-afada84f0505"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.513073 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a3e1b36d-eda4-4ac1-a417-afada84f0505-pod-info" (OuterVolumeSpecName: "pod-info") pod "a3e1b36d-eda4-4ac1-a417-afada84f0505" (UID: "a3e1b36d-eda4-4ac1-a417-afada84f0505"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.519118 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3e1b36d-eda4-4ac1-a417-afada84f0505-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a3e1b36d-eda4-4ac1-a417-afada84f0505" (UID: "a3e1b36d-eda4-4ac1-a417-afada84f0505"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.519331 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-kube-api-access-tcxsd" (OuterVolumeSpecName: "kube-api-access-tcxsd") pod "a3e1b36d-eda4-4ac1-a417-afada84f0505" (UID: "a3e1b36d-eda4-4ac1-a417-afada84f0505"). InnerVolumeSpecName "kube-api-access-tcxsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.520631 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38" (OuterVolumeSpecName: "persistence") pod "a3e1b36d-eda4-4ac1-a417-afada84f0505" (UID: "a3e1b36d-eda4-4ac1-a417-afada84f0505"). InnerVolumeSpecName "pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.530409 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-server-conf" (OuterVolumeSpecName: "server-conf") pod "a3e1b36d-eda4-4ac1-a417-afada84f0505" (UID: "a3e1b36d-eda4-4ac1-a417-afada84f0505"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.593808 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a3e1b36d-eda4-4ac1-a417-afada84f0505" (UID: "a3e1b36d-eda4-4ac1-a417-afada84f0505"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.609116 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.609143 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.609152 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcxsd\" (UniqueName: \"kubernetes.io/projected/a3e1b36d-eda4-4ac1-a417-afada84f0505-kube-api-access-tcxsd\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.609161 4944 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a3e1b36d-eda4-4ac1-a417-afada84f0505-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.609169 4944 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a3e1b36d-eda4-4ac1-a417-afada84f0505-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.609177 4944 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.609210 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\") on node \"crc\" " Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.609220 4944 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a3e1b36d-eda4-4ac1-a417-afada84f0505-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.609230 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a3e1b36d-eda4-4ac1-a417-afada84f0505-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.625790 4944 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.625963 4944 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38") on node "crc" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.631995 4944 generic.go:334] "Generic (PLEG): container finished" podID="a3e1b36d-eda4-4ac1-a417-afada84f0505" containerID="da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14" exitCode=0 Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.632061 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a3e1b36d-eda4-4ac1-a417-afada84f0505","Type":"ContainerDied","Data":"da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14"} Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.632083 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.632112 4944 scope.go:117] "RemoveContainer" containerID="da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.632098 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a3e1b36d-eda4-4ac1-a417-afada84f0505","Type":"ContainerDied","Data":"795ef41bf3c7504dd3edc1231b072a2f39d1d30cab5a4a099b24cdcf0815647d"} Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.651621 4944 scope.go:117] "RemoveContainer" containerID="4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.666620 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.673572 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.694348 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 10:09:11 crc kubenswrapper[4944]: E1124 10:09:11.694658 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3e1b36d-eda4-4ac1-a417-afada84f0505" containerName="setup-container" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.694672 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3e1b36d-eda4-4ac1-a417-afada84f0505" containerName="setup-container" Nov 24 10:09:11 crc kubenswrapper[4944]: E1124 10:09:11.694684 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3e1b36d-eda4-4ac1-a417-afada84f0505" containerName="rabbitmq" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.694690 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3e1b36d-eda4-4ac1-a417-afada84f0505" containerName="rabbitmq" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.694834 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3e1b36d-eda4-4ac1-a417-afada84f0505" containerName="rabbitmq" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.695660 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.698325 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.698495 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-tqqcc" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.698641 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.699460 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.700577 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.711309 4944 reconciler_common.go:293] "Volume detached for volume \"pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.718582 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.763265 4944 scope.go:117] "RemoveContainer" containerID="da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14" Nov 24 10:09:11 crc kubenswrapper[4944]: E1124 10:09:11.763610 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14\": container with ID starting with da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14 not found: ID does not exist" containerID="da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.763646 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14"} err="failed to get container status \"da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14\": rpc error: code = NotFound desc = could not find container \"da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14\": container with ID starting with da582a8316adea7d2b6425d2959bf84e3a2d996373fae37189a3c192ae52cd14 not found: ID does not exist" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.763671 4944 scope.go:117] "RemoveContainer" containerID="4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff" Nov 24 10:09:11 crc kubenswrapper[4944]: E1124 10:09:11.763976 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff\": container with ID starting with 4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff not found: ID does not exist" containerID="4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.764031 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff"} err="failed to get container status \"4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff\": rpc error: code = NotFound desc = could not find container \"4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff\": container with ID starting with 4636c9b582f318dade524c81846e7c54b6c129850bc310dfd1d00658741013ff not found: ID does not exist" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.813100 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d917891-d117-4058-81bd-e582da62b26e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.813345 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d917891-d117-4058-81bd-e582da62b26e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.813402 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d917891-d117-4058-81bd-e582da62b26e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.813436 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d917891-d117-4058-81bd-e582da62b26e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.813455 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfql9\" (UniqueName: \"kubernetes.io/projected/7d917891-d117-4058-81bd-e582da62b26e-kube-api-access-xfql9\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.813500 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.813569 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d917891-d117-4058-81bd-e582da62b26e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.813634 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d917891-d117-4058-81bd-e582da62b26e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.813661 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d917891-d117-4058-81bd-e582da62b26e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.914740 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d917891-d117-4058-81bd-e582da62b26e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.914799 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d917891-d117-4058-81bd-e582da62b26e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.914850 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d917891-d117-4058-81bd-e582da62b26e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.914918 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d917891-d117-4058-81bd-e582da62b26e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.914940 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d917891-d117-4058-81bd-e582da62b26e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.914974 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d917891-d117-4058-81bd-e582da62b26e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.914994 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfql9\" (UniqueName: \"kubernetes.io/projected/7d917891-d117-4058-81bd-e582da62b26e-kube-api-access-xfql9\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.915027 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.915087 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d917891-d117-4058-81bd-e582da62b26e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.916693 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d917891-d117-4058-81bd-e582da62b26e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.917591 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d917891-d117-4058-81bd-e582da62b26e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.917625 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d917891-d117-4058-81bd-e582da62b26e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.917806 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d917891-d117-4058-81bd-e582da62b26e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.919435 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d917891-d117-4058-81bd-e582da62b26e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.919553 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.919574 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/99eac479330c62c6c6b54cc2cb51a2701c71944c1e091963607c33a2179a57cb/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.920708 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d917891-d117-4058-81bd-e582da62b26e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.922123 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d917891-d117-4058-81bd-e582da62b26e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.934016 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfql9\" (UniqueName: \"kubernetes.io/projected/7d917891-d117-4058-81bd-e582da62b26e-kube-api-access-xfql9\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.947456 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-25b16cf4-c8e6-452c-bb14-aedbc2643a38\") pod \"rabbitmq-server-0\" (UID: \"7d917891-d117-4058-81bd-e582da62b26e\") " pod="openstack/rabbitmq-server-0" Nov 24 10:09:11 crc kubenswrapper[4944]: I1124 10:09:11.965112 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.029977 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.119484 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-plugins\") pod \"8248999f-3fee-487c-8c29-a07bf54b8954\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.119824 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8248999f-3fee-487c-8c29-a07bf54b8954-pod-info\") pod \"8248999f-3fee-487c-8c29-a07bf54b8954\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.119883 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dvxz\" (UniqueName: \"kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-kube-api-access-2dvxz\") pod \"8248999f-3fee-487c-8c29-a07bf54b8954\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.119912 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-server-conf\") pod \"8248999f-3fee-487c-8c29-a07bf54b8954\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.119966 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-plugins-conf\") pod \"8248999f-3fee-487c-8c29-a07bf54b8954\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.119986 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-confd\") pod \"8248999f-3fee-487c-8c29-a07bf54b8954\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.120007 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8248999f-3fee-487c-8c29-a07bf54b8954-erlang-cookie-secret\") pod \"8248999f-3fee-487c-8c29-a07bf54b8954\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.120022 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-erlang-cookie\") pod \"8248999f-3fee-487c-8c29-a07bf54b8954\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.120179 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\") pod \"8248999f-3fee-487c-8c29-a07bf54b8954\" (UID: \"8248999f-3fee-487c-8c29-a07bf54b8954\") " Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.121002 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "8248999f-3fee-487c-8c29-a07bf54b8954" (UID: "8248999f-3fee-487c-8c29-a07bf54b8954"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.126174 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "8248999f-3fee-487c-8c29-a07bf54b8954" (UID: "8248999f-3fee-487c-8c29-a07bf54b8954"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.126789 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "8248999f-3fee-487c-8c29-a07bf54b8954" (UID: "8248999f-3fee-487c-8c29-a07bf54b8954"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.141391 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8248999f-3fee-487c-8c29-a07bf54b8954-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "8248999f-3fee-487c-8c29-a07bf54b8954" (UID: "8248999f-3fee-487c-8c29-a07bf54b8954"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.151297 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/8248999f-3fee-487c-8c29-a07bf54b8954-pod-info" (OuterVolumeSpecName: "pod-info") pod "8248999f-3fee-487c-8c29-a07bf54b8954" (UID: "8248999f-3fee-487c-8c29-a07bf54b8954"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.157250 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-kube-api-access-2dvxz" (OuterVolumeSpecName: "kube-api-access-2dvxz") pod "8248999f-3fee-487c-8c29-a07bf54b8954" (UID: "8248999f-3fee-487c-8c29-a07bf54b8954"). InnerVolumeSpecName "kube-api-access-2dvxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.187141 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-server-conf" (OuterVolumeSpecName: "server-conf") pod "8248999f-3fee-487c-8c29-a07bf54b8954" (UID: "8248999f-3fee-487c-8c29-a07bf54b8954"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.227269 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.227302 4944 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8248999f-3fee-487c-8c29-a07bf54b8954-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.227315 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dvxz\" (UniqueName: \"kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-kube-api-access-2dvxz\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.227326 4944 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.227337 4944 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8248999f-3fee-487c-8c29-a07bf54b8954-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.227347 4944 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8248999f-3fee-487c-8c29-a07bf54b8954-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.227358 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.255794 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a" (OuterVolumeSpecName: "persistence") pod "8248999f-3fee-487c-8c29-a07bf54b8954" (UID: "8248999f-3fee-487c-8c29-a07bf54b8954"). InnerVolumeSpecName "pvc-77210e2c-58e3-4d95-ab48-de9f764a840a". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.303178 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3e1b36d-eda4-4ac1-a417-afada84f0505" path="/var/lib/kubelet/pods/a3e1b36d-eda4-4ac1-a417-afada84f0505/volumes" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.307372 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "8248999f-3fee-487c-8c29-a07bf54b8954" (UID: "8248999f-3fee-487c-8c29-a07bf54b8954"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.328915 4944 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8248999f-3fee-487c-8c29-a07bf54b8954-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.328973 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\") on node \"crc\" " Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.371551 4944 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.371908 4944 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-77210e2c-58e3-4d95-ab48-de9f764a840a" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a") on node "crc" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.430704 4944 reconciler_common.go:293] "Volume detached for volume \"pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.630824 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.643366 4944 generic.go:334] "Generic (PLEG): container finished" podID="8248999f-3fee-487c-8c29-a07bf54b8954" containerID="8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536" exitCode=0 Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.643411 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8248999f-3fee-487c-8c29-a07bf54b8954","Type":"ContainerDied","Data":"8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536"} Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.643437 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8248999f-3fee-487c-8c29-a07bf54b8954","Type":"ContainerDied","Data":"456ce97b15f138e75d856e8fbdbfd13b9e8dd82f7459b21b8463832c43e3c5bc"} Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.643436 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.643455 4944 scope.go:117] "RemoveContainer" containerID="8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.670484 4944 scope.go:117] "RemoveContainer" containerID="90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.680867 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.690805 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.712794 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 10:09:12 crc kubenswrapper[4944]: E1124 10:09:12.713358 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8248999f-3fee-487c-8c29-a07bf54b8954" containerName="setup-container" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.713380 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8248999f-3fee-487c-8c29-a07bf54b8954" containerName="setup-container" Nov 24 10:09:12 crc kubenswrapper[4944]: E1124 10:09:12.713428 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8248999f-3fee-487c-8c29-a07bf54b8954" containerName="rabbitmq" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.713436 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8248999f-3fee-487c-8c29-a07bf54b8954" containerName="rabbitmq" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.713618 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8248999f-3fee-487c-8c29-a07bf54b8954" containerName="rabbitmq" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.716212 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.718101 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-q66nf" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.718381 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.718548 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.718711 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.718899 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.724751 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.730541 4944 scope.go:117] "RemoveContainer" containerID="8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536" Nov 24 10:09:12 crc kubenswrapper[4944]: E1124 10:09:12.737360 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536\": container with ID starting with 8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536 not found: ID does not exist" containerID="8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.737420 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536"} err="failed to get container status \"8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536\": rpc error: code = NotFound desc = could not find container \"8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536\": container with ID starting with 8cc4e663009cba17595712af293f171d16622cd3f890125e4ebab1c996ba3536 not found: ID does not exist" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.737452 4944 scope.go:117] "RemoveContainer" containerID="90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90" Nov 24 10:09:12 crc kubenswrapper[4944]: E1124 10:09:12.738599 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90\": container with ID starting with 90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90 not found: ID does not exist" containerID="90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.738654 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90"} err="failed to get container status \"90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90\": rpc error: code = NotFound desc = could not find container \"90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90\": container with ID starting with 90ee46cb4400bfc65e37685b6983e49ece3cdfe81f054ec09346ef6da0d08c90 not found: ID does not exist" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.837863 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.837925 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsmmj\" (UniqueName: \"kubernetes.io/projected/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-kube-api-access-gsmmj\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.838083 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.838145 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.838195 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.838266 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.838355 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.838400 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.838431 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.940178 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.940244 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.940282 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.940285 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.940306 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.940333 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.940349 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.940370 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.940397 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.940432 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsmmj\" (UniqueName: \"kubernetes.io/projected/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-kube-api-access-gsmmj\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.941697 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.942643 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.942817 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.942874 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.942904 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/33b9a91498c17c71a0d96824319b148f89c736699153201088587a2ad180b2b7/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.943104 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.945813 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.945924 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.946504 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:12 crc kubenswrapper[4944]: I1124 10:09:12.962292 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsmmj\" (UniqueName: \"kubernetes.io/projected/15ce2947-cdd4-4b26-99c1-8cbcbacd3905-kube-api-access-gsmmj\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.006956 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-qczm9"] Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.023282 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" podUID="6a9b2075-dce3-4d8c-8ef4-07c885521442" containerName="dnsmasq-dns" containerID="cri-o://da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c" gracePeriod=10 Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.054794 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-77210e2c-58e3-4d95-ab48-de9f764a840a\") pod \"rabbitmq-cell1-server-0\" (UID: \"15ce2947-cdd4-4b26-99c1-8cbcbacd3905\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.335985 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.566269 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.647642 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-config\") pod \"6a9b2075-dce3-4d8c-8ef4-07c885521442\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.647745 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-dns-svc\") pod \"6a9b2075-dce3-4d8c-8ef4-07c885521442\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.647838 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97lwx\" (UniqueName: \"kubernetes.io/projected/6a9b2075-dce3-4d8c-8ef4-07c885521442-kube-api-access-97lwx\") pod \"6a9b2075-dce3-4d8c-8ef4-07c885521442\" (UID: \"6a9b2075-dce3-4d8c-8ef4-07c885521442\") " Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.654419 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d917891-d117-4058-81bd-e582da62b26e","Type":"ContainerStarted","Data":"fa6936878db4de0efc655283d71d062347fe3c8c3ce41c6db0155c018b02e28c"} Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.656260 4944 generic.go:334] "Generic (PLEG): container finished" podID="6a9b2075-dce3-4d8c-8ef4-07c885521442" containerID="da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c" exitCode=0 Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.656311 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" event={"ID":"6a9b2075-dce3-4d8c-8ef4-07c885521442","Type":"ContainerDied","Data":"da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c"} Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.656317 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.656346 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-qczm9" event={"ID":"6a9b2075-dce3-4d8c-8ef4-07c885521442","Type":"ContainerDied","Data":"68cd8314b624e13c779aea254d7ef4d2e4275cf47fc45bf575290e2f7bd1a053"} Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.656374 4944 scope.go:117] "RemoveContainer" containerID="da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.674877 4944 scope.go:117] "RemoveContainer" containerID="bbc96569ad3797a12427562a1b5f34e4f1b124f2e73c480f7efe294cbe56f625" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.695099 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a9b2075-dce3-4d8c-8ef4-07c885521442-kube-api-access-97lwx" (OuterVolumeSpecName: "kube-api-access-97lwx") pod "6a9b2075-dce3-4d8c-8ef4-07c885521442" (UID: "6a9b2075-dce3-4d8c-8ef4-07c885521442"). InnerVolumeSpecName "kube-api-access-97lwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.704230 4944 scope.go:117] "RemoveContainer" containerID="da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c" Nov 24 10:09:13 crc kubenswrapper[4944]: E1124 10:09:13.705059 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c\": container with ID starting with da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c not found: ID does not exist" containerID="da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.705115 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c"} err="failed to get container status \"da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c\": rpc error: code = NotFound desc = could not find container \"da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c\": container with ID starting with da2076617a364fce5e12e7d4ee4c1c2062065f5b01b89fd53d9a28e28ff2835c not found: ID does not exist" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.705149 4944 scope.go:117] "RemoveContainer" containerID="bbc96569ad3797a12427562a1b5f34e4f1b124f2e73c480f7efe294cbe56f625" Nov 24 10:09:13 crc kubenswrapper[4944]: E1124 10:09:13.705854 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbc96569ad3797a12427562a1b5f34e4f1b124f2e73c480f7efe294cbe56f625\": container with ID starting with bbc96569ad3797a12427562a1b5f34e4f1b124f2e73c480f7efe294cbe56f625 not found: ID does not exist" containerID="bbc96569ad3797a12427562a1b5f34e4f1b124f2e73c480f7efe294cbe56f625" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.705921 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbc96569ad3797a12427562a1b5f34e4f1b124f2e73c480f7efe294cbe56f625"} err="failed to get container status \"bbc96569ad3797a12427562a1b5f34e4f1b124f2e73c480f7efe294cbe56f625\": rpc error: code = NotFound desc = could not find container \"bbc96569ad3797a12427562a1b5f34e4f1b124f2e73c480f7efe294cbe56f625\": container with ID starting with bbc96569ad3797a12427562a1b5f34e4f1b124f2e73c480f7efe294cbe56f625 not found: ID does not exist" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.730939 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6a9b2075-dce3-4d8c-8ef4-07c885521442" (UID: "6a9b2075-dce3-4d8c-8ef4-07c885521442"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.740841 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-config" (OuterVolumeSpecName: "config") pod "6a9b2075-dce3-4d8c-8ef4-07c885521442" (UID: "6a9b2075-dce3-4d8c-8ef4-07c885521442"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.750508 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97lwx\" (UniqueName: \"kubernetes.io/projected/6a9b2075-dce3-4d8c-8ef4-07c885521442-kube-api-access-97lwx\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.750544 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.750558 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a9b2075-dce3-4d8c-8ef4-07c885521442-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.988497 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-qczm9"] Nov 24 10:09:13 crc kubenswrapper[4944]: I1124 10:09:13.995260 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-qczm9"] Nov 24 10:09:14 crc kubenswrapper[4944]: I1124 10:09:14.138986 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 10:09:14 crc kubenswrapper[4944]: W1124 10:09:14.140573 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15ce2947_cdd4_4b26_99c1_8cbcbacd3905.slice/crio-239d7697aa6083761d980e98143ece1bc3bb33e982a06ebf366edeb1040a8e0e WatchSource:0}: Error finding container 239d7697aa6083761d980e98143ece1bc3bb33e982a06ebf366edeb1040a8e0e: Status 404 returned error can't find the container with id 239d7697aa6083761d980e98143ece1bc3bb33e982a06ebf366edeb1040a8e0e Nov 24 10:09:14 crc kubenswrapper[4944]: I1124 10:09:14.287838 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a9b2075-dce3-4d8c-8ef4-07c885521442" path="/var/lib/kubelet/pods/6a9b2075-dce3-4d8c-8ef4-07c885521442/volumes" Nov 24 10:09:14 crc kubenswrapper[4944]: I1124 10:09:14.288727 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8248999f-3fee-487c-8c29-a07bf54b8954" path="/var/lib/kubelet/pods/8248999f-3fee-487c-8c29-a07bf54b8954/volumes" Nov 24 10:09:14 crc kubenswrapper[4944]: I1124 10:09:14.664700 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"15ce2947-cdd4-4b26-99c1-8cbcbacd3905","Type":"ContainerStarted","Data":"239d7697aa6083761d980e98143ece1bc3bb33e982a06ebf366edeb1040a8e0e"} Nov 24 10:09:14 crc kubenswrapper[4944]: I1124 10:09:14.668812 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d917891-d117-4058-81bd-e582da62b26e","Type":"ContainerStarted","Data":"2eb8814e294087c4ec17a500407ba6f25d0fbbac19a04451859df75f624d058c"} Nov 24 10:09:15 crc kubenswrapper[4944]: I1124 10:09:15.677156 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"15ce2947-cdd4-4b26-99c1-8cbcbacd3905","Type":"ContainerStarted","Data":"7774d792d3aaca3027932ca1de2c1ac1e450781b3fe58d0110fabc053decd5a0"} Nov 24 10:09:46 crc kubenswrapper[4944]: I1124 10:09:46.895023 4944 generic.go:334] "Generic (PLEG): container finished" podID="7d917891-d117-4058-81bd-e582da62b26e" containerID="2eb8814e294087c4ec17a500407ba6f25d0fbbac19a04451859df75f624d058c" exitCode=0 Nov 24 10:09:46 crc kubenswrapper[4944]: I1124 10:09:46.895128 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d917891-d117-4058-81bd-e582da62b26e","Type":"ContainerDied","Data":"2eb8814e294087c4ec17a500407ba6f25d0fbbac19a04451859df75f624d058c"} Nov 24 10:09:47 crc kubenswrapper[4944]: I1124 10:09:47.903178 4944 generic.go:334] "Generic (PLEG): container finished" podID="15ce2947-cdd4-4b26-99c1-8cbcbacd3905" containerID="7774d792d3aaca3027932ca1de2c1ac1e450781b3fe58d0110fabc053decd5a0" exitCode=0 Nov 24 10:09:47 crc kubenswrapper[4944]: I1124 10:09:47.903285 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"15ce2947-cdd4-4b26-99c1-8cbcbacd3905","Type":"ContainerDied","Data":"7774d792d3aaca3027932ca1de2c1ac1e450781b3fe58d0110fabc053decd5a0"} Nov 24 10:09:47 crc kubenswrapper[4944]: I1124 10:09:47.906354 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d917891-d117-4058-81bd-e582da62b26e","Type":"ContainerStarted","Data":"01c2447efaac75a1ac72d9a1b60dea10629551e05dc48dcc88d541c932355510"} Nov 24 10:09:47 crc kubenswrapper[4944]: I1124 10:09:47.906593 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 10:09:47 crc kubenswrapper[4944]: I1124 10:09:47.982105 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.982085187 podStartE2EDuration="36.982085187s" podCreationTimestamp="2025-11-24 10:09:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:09:47.979809994 +0000 UTC m=+4648.514250466" watchObservedRunningTime="2025-11-24 10:09:47.982085187 +0000 UTC m=+4648.516525649" Nov 24 10:09:48 crc kubenswrapper[4944]: I1124 10:09:48.915556 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"15ce2947-cdd4-4b26-99c1-8cbcbacd3905","Type":"ContainerStarted","Data":"6a77f433f76a17ab4f21d31ae38d2c79716793b9a033e417c6cf4c16994be6dc"} Nov 24 10:09:48 crc kubenswrapper[4944]: I1124 10:09:48.916113 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:09:48 crc kubenswrapper[4944]: I1124 10:09:48.936272 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.936254393 podStartE2EDuration="36.936254393s" podCreationTimestamp="2025-11-24 10:09:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:09:48.932208724 +0000 UTC m=+4649.466649186" watchObservedRunningTime="2025-11-24 10:09:48.936254393 +0000 UTC m=+4649.470694855" Nov 24 10:10:02 crc kubenswrapper[4944]: I1124 10:10:02.033130 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 10:10:03 crc kubenswrapper[4944]: I1124 10:10:03.340581 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 10:10:03 crc kubenswrapper[4944]: I1124 10:10:03.818381 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lmbvd"] Nov 24 10:10:03 crc kubenswrapper[4944]: E1124 10:10:03.819021 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a9b2075-dce3-4d8c-8ef4-07c885521442" containerName="dnsmasq-dns" Nov 24 10:10:03 crc kubenswrapper[4944]: I1124 10:10:03.819123 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a9b2075-dce3-4d8c-8ef4-07c885521442" containerName="dnsmasq-dns" Nov 24 10:10:03 crc kubenswrapper[4944]: E1124 10:10:03.819198 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a9b2075-dce3-4d8c-8ef4-07c885521442" containerName="init" Nov 24 10:10:03 crc kubenswrapper[4944]: I1124 10:10:03.819261 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a9b2075-dce3-4d8c-8ef4-07c885521442" containerName="init" Nov 24 10:10:03 crc kubenswrapper[4944]: I1124 10:10:03.819522 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a9b2075-dce3-4d8c-8ef4-07c885521442" containerName="dnsmasq-dns" Nov 24 10:10:03 crc kubenswrapper[4944]: I1124 10:10:03.820857 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:03 crc kubenswrapper[4944]: I1124 10:10:03.834513 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lmbvd"] Nov 24 10:10:03 crc kubenswrapper[4944]: I1124 10:10:03.900096 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t4qq\" (UniqueName: \"kubernetes.io/projected/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-kube-api-access-8t4qq\") pod \"redhat-operators-lmbvd\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:03 crc kubenswrapper[4944]: I1124 10:10:03.900173 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-catalog-content\") pod \"redhat-operators-lmbvd\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:03 crc kubenswrapper[4944]: I1124 10:10:03.900251 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-utilities\") pod \"redhat-operators-lmbvd\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:04 crc kubenswrapper[4944]: I1124 10:10:04.001213 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-catalog-content\") pod \"redhat-operators-lmbvd\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:04 crc kubenswrapper[4944]: I1124 10:10:04.001275 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-utilities\") pod \"redhat-operators-lmbvd\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:04 crc kubenswrapper[4944]: I1124 10:10:04.001347 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t4qq\" (UniqueName: \"kubernetes.io/projected/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-kube-api-access-8t4qq\") pod \"redhat-operators-lmbvd\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:04 crc kubenswrapper[4944]: I1124 10:10:04.001758 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-catalog-content\") pod \"redhat-operators-lmbvd\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:04 crc kubenswrapper[4944]: I1124 10:10:04.001843 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-utilities\") pod \"redhat-operators-lmbvd\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:04 crc kubenswrapper[4944]: I1124 10:10:04.024940 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t4qq\" (UniqueName: \"kubernetes.io/projected/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-kube-api-access-8t4qq\") pod \"redhat-operators-lmbvd\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:04 crc kubenswrapper[4944]: I1124 10:10:04.147741 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:04 crc kubenswrapper[4944]: I1124 10:10:04.569044 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lmbvd"] Nov 24 10:10:05 crc kubenswrapper[4944]: I1124 10:10:05.021570 4944 generic.go:334] "Generic (PLEG): container finished" podID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" containerID="e274ae13a1053ce234bbcefa6d2d2a30820f973ce6e5f43ceb02322888531b50" exitCode=0 Nov 24 10:10:05 crc kubenswrapper[4944]: I1124 10:10:05.021621 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmbvd" event={"ID":"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24","Type":"ContainerDied","Data":"e274ae13a1053ce234bbcefa6d2d2a30820f973ce6e5f43ceb02322888531b50"} Nov 24 10:10:05 crc kubenswrapper[4944]: I1124 10:10:05.021647 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmbvd" event={"ID":"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24","Type":"ContainerStarted","Data":"61d534ef1d0816e6e95e8b2a27f5a7087acb5c36a6d19ff211a3c877e98c76a5"} Nov 24 10:10:05 crc kubenswrapper[4944]: I1124 10:10:05.024288 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 10:10:06 crc kubenswrapper[4944]: I1124 10:10:06.029032 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmbvd" event={"ID":"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24","Type":"ContainerStarted","Data":"3e888414625b380a55ae354c3119b53b1a7bfe1569c8badd85669990eafb52af"} Nov 24 10:10:07 crc kubenswrapper[4944]: I1124 10:10:07.038491 4944 generic.go:334] "Generic (PLEG): container finished" podID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" containerID="3e888414625b380a55ae354c3119b53b1a7bfe1569c8badd85669990eafb52af" exitCode=0 Nov 24 10:10:07 crc kubenswrapper[4944]: I1124 10:10:07.038534 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmbvd" event={"ID":"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24","Type":"ContainerDied","Data":"3e888414625b380a55ae354c3119b53b1a7bfe1569c8badd85669990eafb52af"} Nov 24 10:10:08 crc kubenswrapper[4944]: I1124 10:10:08.050155 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmbvd" event={"ID":"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24","Type":"ContainerStarted","Data":"a55fbc8f35779d42348ddca743f1be257ec76ad9fa751f47de40688d040d2f71"} Nov 24 10:10:08 crc kubenswrapper[4944]: I1124 10:10:08.078011 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lmbvd" podStartSLOduration=2.594950743 podStartE2EDuration="5.077991846s" podCreationTimestamp="2025-11-24 10:10:03 +0000 UTC" firstStartedPulling="2025-11-24 10:10:05.024051083 +0000 UTC m=+4665.558491545" lastFinishedPulling="2025-11-24 10:10:07.507092186 +0000 UTC m=+4668.041532648" observedRunningTime="2025-11-24 10:10:08.073740721 +0000 UTC m=+4668.608181203" watchObservedRunningTime="2025-11-24 10:10:08.077991846 +0000 UTC m=+4668.612432298" Nov 24 10:10:14 crc kubenswrapper[4944]: I1124 10:10:14.148465 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:14 crc kubenswrapper[4944]: I1124 10:10:14.148984 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:14 crc kubenswrapper[4944]: I1124 10:10:14.199474 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:14 crc kubenswrapper[4944]: I1124 10:10:14.478477 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 10:10:14 crc kubenswrapper[4944]: I1124 10:10:14.479761 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 10:10:14 crc kubenswrapper[4944]: I1124 10:10:14.481928 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-v4r8j" Nov 24 10:10:14 crc kubenswrapper[4944]: I1124 10:10:14.486457 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 10:10:14 crc kubenswrapper[4944]: I1124 10:10:14.550762 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hsj2\" (UniqueName: \"kubernetes.io/projected/c00bab8b-29ee-4e84-8386-4ec77fe99f31-kube-api-access-8hsj2\") pod \"mariadb-client-1-default\" (UID: \"c00bab8b-29ee-4e84-8386-4ec77fe99f31\") " pod="openstack/mariadb-client-1-default" Nov 24 10:10:14 crc kubenswrapper[4944]: I1124 10:10:14.652388 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hsj2\" (UniqueName: \"kubernetes.io/projected/c00bab8b-29ee-4e84-8386-4ec77fe99f31-kube-api-access-8hsj2\") pod \"mariadb-client-1-default\" (UID: \"c00bab8b-29ee-4e84-8386-4ec77fe99f31\") " pod="openstack/mariadb-client-1-default" Nov 24 10:10:14 crc kubenswrapper[4944]: I1124 10:10:14.686127 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hsj2\" (UniqueName: \"kubernetes.io/projected/c00bab8b-29ee-4e84-8386-4ec77fe99f31-kube-api-access-8hsj2\") pod \"mariadb-client-1-default\" (UID: \"c00bab8b-29ee-4e84-8386-4ec77fe99f31\") " pod="openstack/mariadb-client-1-default" Nov 24 10:10:14 crc kubenswrapper[4944]: I1124 10:10:14.799510 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 10:10:15 crc kubenswrapper[4944]: I1124 10:10:15.134238 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:15 crc kubenswrapper[4944]: I1124 10:10:15.192519 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lmbvd"] Nov 24 10:10:15 crc kubenswrapper[4944]: I1124 10:10:15.382768 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 10:10:16 crc kubenswrapper[4944]: I1124 10:10:16.102890 4944 generic.go:334] "Generic (PLEG): container finished" podID="c00bab8b-29ee-4e84-8386-4ec77fe99f31" containerID="d1d5a9d00c4e83b11041f4eb063e3b07b6cfb4aca9171c38c3c98b45a9fb10d8" exitCode=0 Nov 24 10:10:16 crc kubenswrapper[4944]: I1124 10:10:16.102941 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"c00bab8b-29ee-4e84-8386-4ec77fe99f31","Type":"ContainerDied","Data":"d1d5a9d00c4e83b11041f4eb063e3b07b6cfb4aca9171c38c3c98b45a9fb10d8"} Nov 24 10:10:16 crc kubenswrapper[4944]: I1124 10:10:16.103237 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"c00bab8b-29ee-4e84-8386-4ec77fe99f31","Type":"ContainerStarted","Data":"12b64c64bf23fe7a25148ff5b711c41be921713acb5ce382f7c483c94ef85d74"} Nov 24 10:10:16 crc kubenswrapper[4944]: I1124 10:10:16.840242 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4lq9j"] Nov 24 10:10:16 crc kubenswrapper[4944]: I1124 10:10:16.842521 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:16 crc kubenswrapper[4944]: I1124 10:10:16.851696 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lq9j"] Nov 24 10:10:16 crc kubenswrapper[4944]: I1124 10:10:16.902708 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-catalog-content\") pod \"redhat-marketplace-4lq9j\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:16 crc kubenswrapper[4944]: I1124 10:10:16.902745 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-utilities\") pod \"redhat-marketplace-4lq9j\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:16 crc kubenswrapper[4944]: I1124 10:10:16.902822 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsmjc\" (UniqueName: \"kubernetes.io/projected/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-kube-api-access-zsmjc\") pod \"redhat-marketplace-4lq9j\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.004032 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsmjc\" (UniqueName: \"kubernetes.io/projected/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-kube-api-access-zsmjc\") pod \"redhat-marketplace-4lq9j\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.004173 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-catalog-content\") pod \"redhat-marketplace-4lq9j\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.004201 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-utilities\") pod \"redhat-marketplace-4lq9j\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.004727 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-utilities\") pod \"redhat-marketplace-4lq9j\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.005009 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-catalog-content\") pod \"redhat-marketplace-4lq9j\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.023582 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsmjc\" (UniqueName: \"kubernetes.io/projected/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-kube-api-access-zsmjc\") pod \"redhat-marketplace-4lq9j\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.109711 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lmbvd" podUID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" containerName="registry-server" containerID="cri-o://a55fbc8f35779d42348ddca743f1be257ec76ad9fa751f47de40688d040d2f71" gracePeriod=2 Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.165301 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.571979 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.600642 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_c00bab8b-29ee-4e84-8386-4ec77fe99f31/mariadb-client-1-default/0.log" Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.611102 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lq9j"] Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.627980 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.635448 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.713876 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hsj2\" (UniqueName: \"kubernetes.io/projected/c00bab8b-29ee-4e84-8386-4ec77fe99f31-kube-api-access-8hsj2\") pod \"c00bab8b-29ee-4e84-8386-4ec77fe99f31\" (UID: \"c00bab8b-29ee-4e84-8386-4ec77fe99f31\") " Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.721425 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c00bab8b-29ee-4e84-8386-4ec77fe99f31-kube-api-access-8hsj2" (OuterVolumeSpecName: "kube-api-access-8hsj2") pod "c00bab8b-29ee-4e84-8386-4ec77fe99f31" (UID: "c00bab8b-29ee-4e84-8386-4ec77fe99f31"). InnerVolumeSpecName "kube-api-access-8hsj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:10:17 crc kubenswrapper[4944]: I1124 10:10:17.816020 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hsj2\" (UniqueName: \"kubernetes.io/projected/c00bab8b-29ee-4e84-8386-4ec77fe99f31-kube-api-access-8hsj2\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.078123 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 10:10:18 crc kubenswrapper[4944]: E1124 10:10:18.078448 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c00bab8b-29ee-4e84-8386-4ec77fe99f31" containerName="mariadb-client-1-default" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.078461 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c00bab8b-29ee-4e84-8386-4ec77fe99f31" containerName="mariadb-client-1-default" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.078602 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="c00bab8b-29ee-4e84-8386-4ec77fe99f31" containerName="mariadb-client-1-default" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.079171 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.089115 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.117221 4944 generic.go:334] "Generic (PLEG): container finished" podID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" containerID="a55fbc8f35779d42348ddca743f1be257ec76ad9fa751f47de40688d040d2f71" exitCode=0 Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.117295 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmbvd" event={"ID":"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24","Type":"ContainerDied","Data":"a55fbc8f35779d42348ddca743f1be257ec76ad9fa751f47de40688d040d2f71"} Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.118733 4944 generic.go:334] "Generic (PLEG): container finished" podID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" containerID="28ba6ef9c3cb23fcfbbbba19deb29f4033668d5b2720dd1b22322760b6619783" exitCode=0 Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.118785 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lq9j" event={"ID":"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e","Type":"ContainerDied","Data":"28ba6ef9c3cb23fcfbbbba19deb29f4033668d5b2720dd1b22322760b6619783"} Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.118805 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lq9j" event={"ID":"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e","Type":"ContainerStarted","Data":"36dc29dc1502acefd7ae3ba88085b02883d8d1f4a20a1127096d7ec23d7ca3a4"} Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.121166 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12b64c64bf23fe7a25148ff5b711c41be921713acb5ce382f7c483c94ef85d74" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.121214 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.221534 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpthj\" (UniqueName: \"kubernetes.io/projected/286eb206-6127-46ec-a3f3-7f91c512e351-kube-api-access-dpthj\") pod \"mariadb-client-2-default\" (UID: \"286eb206-6127-46ec-a3f3-7f91c512e351\") " pod="openstack/mariadb-client-2-default" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.286513 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c00bab8b-29ee-4e84-8386-4ec77fe99f31" path="/var/lib/kubelet/pods/c00bab8b-29ee-4e84-8386-4ec77fe99f31/volumes" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.323630 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpthj\" (UniqueName: \"kubernetes.io/projected/286eb206-6127-46ec-a3f3-7f91c512e351-kube-api-access-dpthj\") pod \"mariadb-client-2-default\" (UID: \"286eb206-6127-46ec-a3f3-7f91c512e351\") " pod="openstack/mariadb-client-2-default" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.339821 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpthj\" (UniqueName: \"kubernetes.io/projected/286eb206-6127-46ec-a3f3-7f91c512e351-kube-api-access-dpthj\") pod \"mariadb-client-2-default\" (UID: \"286eb206-6127-46ec-a3f3-7f91c512e351\") " pod="openstack/mariadb-client-2-default" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.399812 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.590827 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.713009 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 10:10:18 crc kubenswrapper[4944]: W1124 10:10:18.717730 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod286eb206_6127_46ec_a3f3_7f91c512e351.slice/crio-85e8953aefcef83ae30704eb205b55de3947e4f547eaa4f920809837c70c9228 WatchSource:0}: Error finding container 85e8953aefcef83ae30704eb205b55de3947e4f547eaa4f920809837c70c9228: Status 404 returned error can't find the container with id 85e8953aefcef83ae30704eb205b55de3947e4f547eaa4f920809837c70c9228 Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.738470 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t4qq\" (UniqueName: \"kubernetes.io/projected/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-kube-api-access-8t4qq\") pod \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.738576 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-catalog-content\") pod \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.738678 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-utilities\") pod \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\" (UID: \"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24\") " Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.739770 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-utilities" (OuterVolumeSpecName: "utilities") pod "be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" (UID: "be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.743641 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-kube-api-access-8t4qq" (OuterVolumeSpecName: "kube-api-access-8t4qq") pod "be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" (UID: "be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24"). InnerVolumeSpecName "kube-api-access-8t4qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.819867 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" (UID: "be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.840668 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.840704 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t4qq\" (UniqueName: \"kubernetes.io/projected/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-kube-api-access-8t4qq\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:18 crc kubenswrapper[4944]: I1124 10:10:18.840718 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:19 crc kubenswrapper[4944]: I1124 10:10:19.131531 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lmbvd" event={"ID":"be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24","Type":"ContainerDied","Data":"61d534ef1d0816e6e95e8b2a27f5a7087acb5c36a6d19ff211a3c877e98c76a5"} Nov 24 10:10:19 crc kubenswrapper[4944]: I1124 10:10:19.131566 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lmbvd" Nov 24 10:10:19 crc kubenswrapper[4944]: I1124 10:10:19.131579 4944 scope.go:117] "RemoveContainer" containerID="a55fbc8f35779d42348ddca743f1be257ec76ad9fa751f47de40688d040d2f71" Nov 24 10:10:19 crc kubenswrapper[4944]: I1124 10:10:19.134884 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"286eb206-6127-46ec-a3f3-7f91c512e351","Type":"ContainerStarted","Data":"ad0b7bc15faa435934ff6851c28afc33050316559d83d902e8b8397ef76e06f7"} Nov 24 10:10:19 crc kubenswrapper[4944]: I1124 10:10:19.134908 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"286eb206-6127-46ec-a3f3-7f91c512e351","Type":"ContainerStarted","Data":"85e8953aefcef83ae30704eb205b55de3947e4f547eaa4f920809837c70c9228"} Nov 24 10:10:19 crc kubenswrapper[4944]: I1124 10:10:19.150087 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=1.150070089 podStartE2EDuration="1.150070089s" podCreationTimestamp="2025-11-24 10:10:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:10:19.146708713 +0000 UTC m=+4679.681149185" watchObservedRunningTime="2025-11-24 10:10:19.150070089 +0000 UTC m=+4679.684510551" Nov 24 10:10:19 crc kubenswrapper[4944]: I1124 10:10:19.168004 4944 scope.go:117] "RemoveContainer" containerID="3e888414625b380a55ae354c3119b53b1a7bfe1569c8badd85669990eafb52af" Nov 24 10:10:19 crc kubenswrapper[4944]: I1124 10:10:19.182702 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lmbvd"] Nov 24 10:10:19 crc kubenswrapper[4944]: I1124 10:10:19.197295 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lmbvd"] Nov 24 10:10:19 crc kubenswrapper[4944]: I1124 10:10:19.204563 4944 scope.go:117] "RemoveContainer" containerID="e274ae13a1053ce234bbcefa6d2d2a30820f973ce6e5f43ceb02322888531b50" Nov 24 10:10:20 crc kubenswrapper[4944]: I1124 10:10:20.145170 4944 generic.go:334] "Generic (PLEG): container finished" podID="286eb206-6127-46ec-a3f3-7f91c512e351" containerID="ad0b7bc15faa435934ff6851c28afc33050316559d83d902e8b8397ef76e06f7" exitCode=1 Nov 24 10:10:20 crc kubenswrapper[4944]: I1124 10:10:20.145252 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"286eb206-6127-46ec-a3f3-7f91c512e351","Type":"ContainerDied","Data":"ad0b7bc15faa435934ff6851c28afc33050316559d83d902e8b8397ef76e06f7"} Nov 24 10:10:20 crc kubenswrapper[4944]: I1124 10:10:20.147934 4944 generic.go:334] "Generic (PLEG): container finished" podID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" containerID="062fa447f1d00122f7151c702827e667fd88d7146e252beac4bbcd197b32bd03" exitCode=0 Nov 24 10:10:20 crc kubenswrapper[4944]: I1124 10:10:20.147971 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lq9j" event={"ID":"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e","Type":"ContainerDied","Data":"062fa447f1d00122f7151c702827e667fd88d7146e252beac4bbcd197b32bd03"} Nov 24 10:10:20 crc kubenswrapper[4944]: I1124 10:10:20.296433 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" path="/var/lib/kubelet/pods/be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24/volumes" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.158354 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lq9j" event={"ID":"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e","Type":"ContainerStarted","Data":"8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61"} Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.179704 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4lq9j" podStartSLOduration=2.760760821 podStartE2EDuration="5.179677832s" podCreationTimestamp="2025-11-24 10:10:16 +0000 UTC" firstStartedPulling="2025-11-24 10:10:18.121636339 +0000 UTC m=+4678.656076801" lastFinishedPulling="2025-11-24 10:10:20.54055335 +0000 UTC m=+4681.074993812" observedRunningTime="2025-11-24 10:10:21.176294255 +0000 UTC m=+4681.710734717" watchObservedRunningTime="2025-11-24 10:10:21.179677832 +0000 UTC m=+4681.714118334" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.504274 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.546478 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.552623 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.581928 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpthj\" (UniqueName: \"kubernetes.io/projected/286eb206-6127-46ec-a3f3-7f91c512e351-kube-api-access-dpthj\") pod \"286eb206-6127-46ec-a3f3-7f91c512e351\" (UID: \"286eb206-6127-46ec-a3f3-7f91c512e351\") " Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.589416 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286eb206-6127-46ec-a3f3-7f91c512e351-kube-api-access-dpthj" (OuterVolumeSpecName: "kube-api-access-dpthj") pod "286eb206-6127-46ec-a3f3-7f91c512e351" (UID: "286eb206-6127-46ec-a3f3-7f91c512e351"). InnerVolumeSpecName "kube-api-access-dpthj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.683725 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpthj\" (UniqueName: \"kubernetes.io/projected/286eb206-6127-46ec-a3f3-7f91c512e351-kube-api-access-dpthj\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.991653 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Nov 24 10:10:21 crc kubenswrapper[4944]: E1124 10:10:21.992009 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" containerName="extract-content" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.992022 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" containerName="extract-content" Nov 24 10:10:21 crc kubenswrapper[4944]: E1124 10:10:21.992039 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" containerName="extract-utilities" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.992073 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" containerName="extract-utilities" Nov 24 10:10:21 crc kubenswrapper[4944]: E1124 10:10:21.992102 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" containerName="registry-server" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.992111 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" containerName="registry-server" Nov 24 10:10:21 crc kubenswrapper[4944]: E1124 10:10:21.992126 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286eb206-6127-46ec-a3f3-7f91c512e351" containerName="mariadb-client-2-default" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.992132 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="286eb206-6127-46ec-a3f3-7f91c512e351" containerName="mariadb-client-2-default" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.992288 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="286eb206-6127-46ec-a3f3-7f91c512e351" containerName="mariadb-client-2-default" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.992301 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="be9cd4a2-a8cd-4a4e-a0b4-025939b0ff24" containerName="registry-server" Nov 24 10:10:21 crc kubenswrapper[4944]: I1124 10:10:21.993196 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 10:10:22 crc kubenswrapper[4944]: I1124 10:10:22.000750 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 10:10:22 crc kubenswrapper[4944]: I1124 10:10:22.089612 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jprhr\" (UniqueName: \"kubernetes.io/projected/8ccd963c-d896-479e-8478-70225a380d6f-kube-api-access-jprhr\") pod \"mariadb-client-1\" (UID: \"8ccd963c-d896-479e-8478-70225a380d6f\") " pod="openstack/mariadb-client-1" Nov 24 10:10:22 crc kubenswrapper[4944]: I1124 10:10:22.166729 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85e8953aefcef83ae30704eb205b55de3947e4f547eaa4f920809837c70c9228" Nov 24 10:10:22 crc kubenswrapper[4944]: I1124 10:10:22.166784 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 24 10:10:22 crc kubenswrapper[4944]: I1124 10:10:22.190462 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jprhr\" (UniqueName: \"kubernetes.io/projected/8ccd963c-d896-479e-8478-70225a380d6f-kube-api-access-jprhr\") pod \"mariadb-client-1\" (UID: \"8ccd963c-d896-479e-8478-70225a380d6f\") " pod="openstack/mariadb-client-1" Nov 24 10:10:22 crc kubenswrapper[4944]: I1124 10:10:22.208950 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jprhr\" (UniqueName: \"kubernetes.io/projected/8ccd963c-d896-479e-8478-70225a380d6f-kube-api-access-jprhr\") pod \"mariadb-client-1\" (UID: \"8ccd963c-d896-479e-8478-70225a380d6f\") " pod="openstack/mariadb-client-1" Nov 24 10:10:22 crc kubenswrapper[4944]: I1124 10:10:22.285416 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="286eb206-6127-46ec-a3f3-7f91c512e351" path="/var/lib/kubelet/pods/286eb206-6127-46ec-a3f3-7f91c512e351/volumes" Nov 24 10:10:22 crc kubenswrapper[4944]: I1124 10:10:22.309758 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 10:10:22 crc kubenswrapper[4944]: E1124 10:10:22.333875 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod286eb206_6127_46ec_a3f3_7f91c512e351.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod286eb206_6127_46ec_a3f3_7f91c512e351.slice/crio-85e8953aefcef83ae30704eb205b55de3947e4f547eaa4f920809837c70c9228\": RecentStats: unable to find data in memory cache]" Nov 24 10:10:22 crc kubenswrapper[4944]: I1124 10:10:22.765715 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 10:10:22 crc kubenswrapper[4944]: W1124 10:10:22.769632 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ccd963c_d896_479e_8478_70225a380d6f.slice/crio-9a3cd4c980b3b263edef84ed3f0a7e4798773e4f757cbbe300f1e48702098d2a WatchSource:0}: Error finding container 9a3cd4c980b3b263edef84ed3f0a7e4798773e4f757cbbe300f1e48702098d2a: Status 404 returned error can't find the container with id 9a3cd4c980b3b263edef84ed3f0a7e4798773e4f757cbbe300f1e48702098d2a Nov 24 10:10:23 crc kubenswrapper[4944]: I1124 10:10:23.177245 4944 generic.go:334] "Generic (PLEG): container finished" podID="8ccd963c-d896-479e-8478-70225a380d6f" containerID="aa2726ee6bfd4f4cf805453a184118501cd210493f886d7d77138f780d87753f" exitCode=0 Nov 24 10:10:23 crc kubenswrapper[4944]: I1124 10:10:23.177348 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"8ccd963c-d896-479e-8478-70225a380d6f","Type":"ContainerDied","Data":"aa2726ee6bfd4f4cf805453a184118501cd210493f886d7d77138f780d87753f"} Nov 24 10:10:23 crc kubenswrapper[4944]: I1124 10:10:23.177561 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"8ccd963c-d896-479e-8478-70225a380d6f","Type":"ContainerStarted","Data":"9a3cd4c980b3b263edef84ed3f0a7e4798773e4f757cbbe300f1e48702098d2a"} Nov 24 10:10:24 crc kubenswrapper[4944]: I1124 10:10:24.555849 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 10:10:24 crc kubenswrapper[4944]: I1124 10:10:24.575567 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_8ccd963c-d896-479e-8478-70225a380d6f/mariadb-client-1/0.log" Nov 24 10:10:24 crc kubenswrapper[4944]: I1124 10:10:24.599975 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 10:10:24 crc kubenswrapper[4944]: I1124 10:10:24.604254 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Nov 24 10:10:24 crc kubenswrapper[4944]: I1124 10:10:24.639994 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jprhr\" (UniqueName: \"kubernetes.io/projected/8ccd963c-d896-479e-8478-70225a380d6f-kube-api-access-jprhr\") pod \"8ccd963c-d896-479e-8478-70225a380d6f\" (UID: \"8ccd963c-d896-479e-8478-70225a380d6f\") " Nov 24 10:10:24 crc kubenswrapper[4944]: I1124 10:10:24.646521 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ccd963c-d896-479e-8478-70225a380d6f-kube-api-access-jprhr" (OuterVolumeSpecName: "kube-api-access-jprhr") pod "8ccd963c-d896-479e-8478-70225a380d6f" (UID: "8ccd963c-d896-479e-8478-70225a380d6f"). InnerVolumeSpecName "kube-api-access-jprhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:10:24 crc kubenswrapper[4944]: I1124 10:10:24.741455 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jprhr\" (UniqueName: \"kubernetes.io/projected/8ccd963c-d896-479e-8478-70225a380d6f-kube-api-access-jprhr\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.019545 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 10:10:25 crc kubenswrapper[4944]: E1124 10:10:25.019989 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ccd963c-d896-479e-8478-70225a380d6f" containerName="mariadb-client-1" Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.020008 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ccd963c-d896-479e-8478-70225a380d6f" containerName="mariadb-client-1" Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.020226 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ccd963c-d896-479e-8478-70225a380d6f" containerName="mariadb-client-1" Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.020887 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.029984 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.146939 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w5fm\" (UniqueName: \"kubernetes.io/projected/79adaf85-ae0b-447e-a908-a78b796edc12-kube-api-access-2w5fm\") pod \"mariadb-client-4-default\" (UID: \"79adaf85-ae0b-447e-a908-a78b796edc12\") " pod="openstack/mariadb-client-4-default" Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.198402 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a3cd4c980b3b263edef84ed3f0a7e4798773e4f757cbbe300f1e48702098d2a" Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.198469 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.249253 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w5fm\" (UniqueName: \"kubernetes.io/projected/79adaf85-ae0b-447e-a908-a78b796edc12-kube-api-access-2w5fm\") pod \"mariadb-client-4-default\" (UID: \"79adaf85-ae0b-447e-a908-a78b796edc12\") " pod="openstack/mariadb-client-4-default" Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.268298 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w5fm\" (UniqueName: \"kubernetes.io/projected/79adaf85-ae0b-447e-a908-a78b796edc12-kube-api-access-2w5fm\") pod \"mariadb-client-4-default\" (UID: \"79adaf85-ae0b-447e-a908-a78b796edc12\") " pod="openstack/mariadb-client-4-default" Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.343738 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 10:10:25 crc kubenswrapper[4944]: I1124 10:10:25.838186 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 10:10:26 crc kubenswrapper[4944]: I1124 10:10:26.207301 4944 generic.go:334] "Generic (PLEG): container finished" podID="79adaf85-ae0b-447e-a908-a78b796edc12" containerID="e9fc5f229012c10e7ad3821d40f878fef079197c3588dca3eba0338022f1480e" exitCode=0 Nov 24 10:10:26 crc kubenswrapper[4944]: I1124 10:10:26.207355 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"79adaf85-ae0b-447e-a908-a78b796edc12","Type":"ContainerDied","Data":"e9fc5f229012c10e7ad3821d40f878fef079197c3588dca3eba0338022f1480e"} Nov 24 10:10:26 crc kubenswrapper[4944]: I1124 10:10:26.207400 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"79adaf85-ae0b-447e-a908-a78b796edc12","Type":"ContainerStarted","Data":"de189008d186fc15ab698b14a3e295ab80a7798984491fa3ee1e7438385ee93a"} Nov 24 10:10:26 crc kubenswrapper[4944]: I1124 10:10:26.284649 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ccd963c-d896-479e-8478-70225a380d6f" path="/var/lib/kubelet/pods/8ccd963c-d896-479e-8478-70225a380d6f/volumes" Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.165820 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.166226 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.237291 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.295877 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.473807 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lq9j"] Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.592290 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.609154 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_79adaf85-ae0b-447e-a908-a78b796edc12/mariadb-client-4-default/0.log" Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.634397 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.639475 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.688246 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w5fm\" (UniqueName: \"kubernetes.io/projected/79adaf85-ae0b-447e-a908-a78b796edc12-kube-api-access-2w5fm\") pod \"79adaf85-ae0b-447e-a908-a78b796edc12\" (UID: \"79adaf85-ae0b-447e-a908-a78b796edc12\") " Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.693772 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79adaf85-ae0b-447e-a908-a78b796edc12-kube-api-access-2w5fm" (OuterVolumeSpecName: "kube-api-access-2w5fm") pod "79adaf85-ae0b-447e-a908-a78b796edc12" (UID: "79adaf85-ae0b-447e-a908-a78b796edc12"). InnerVolumeSpecName "kube-api-access-2w5fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:10:27 crc kubenswrapper[4944]: I1124 10:10:27.790679 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w5fm\" (UniqueName: \"kubernetes.io/projected/79adaf85-ae0b-447e-a908-a78b796edc12-kube-api-access-2w5fm\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:28 crc kubenswrapper[4944]: I1124 10:10:28.224787 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de189008d186fc15ab698b14a3e295ab80a7798984491fa3ee1e7438385ee93a" Nov 24 10:10:28 crc kubenswrapper[4944]: I1124 10:10:28.225392 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 24 10:10:28 crc kubenswrapper[4944]: I1124 10:10:28.289431 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79adaf85-ae0b-447e-a908-a78b796edc12" path="/var/lib/kubelet/pods/79adaf85-ae0b-447e-a908-a78b796edc12/volumes" Nov 24 10:10:29 crc kubenswrapper[4944]: I1124 10:10:29.232697 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4lq9j" podUID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" containerName="registry-server" containerID="cri-o://8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61" gracePeriod=2 Nov 24 10:10:29 crc kubenswrapper[4944]: I1124 10:10:29.617772 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:29 crc kubenswrapper[4944]: I1124 10:10:29.730928 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-utilities\") pod \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " Nov 24 10:10:29 crc kubenswrapper[4944]: I1124 10:10:29.730989 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-catalog-content\") pod \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " Nov 24 10:10:29 crc kubenswrapper[4944]: I1124 10:10:29.731019 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsmjc\" (UniqueName: \"kubernetes.io/projected/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-kube-api-access-zsmjc\") pod \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\" (UID: \"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e\") " Nov 24 10:10:29 crc kubenswrapper[4944]: I1124 10:10:29.734998 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-utilities" (OuterVolumeSpecName: "utilities") pod "0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" (UID: "0c08ee78-66ee-4c9a-b779-ab667bdb8e7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:10:29 crc kubenswrapper[4944]: I1124 10:10:29.750810 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-kube-api-access-zsmjc" (OuterVolumeSpecName: "kube-api-access-zsmjc") pod "0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" (UID: "0c08ee78-66ee-4c9a-b779-ab667bdb8e7e"). InnerVolumeSpecName "kube-api-access-zsmjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:10:29 crc kubenswrapper[4944]: I1124 10:10:29.766452 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" (UID: "0c08ee78-66ee-4c9a-b779-ab667bdb8e7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:10:29 crc kubenswrapper[4944]: I1124 10:10:29.832402 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:29 crc kubenswrapper[4944]: I1124 10:10:29.832440 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:29 crc kubenswrapper[4944]: I1124 10:10:29.832454 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsmjc\" (UniqueName: \"kubernetes.io/projected/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e-kube-api-access-zsmjc\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.240954 4944 generic.go:334] "Generic (PLEG): container finished" podID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" containerID="8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61" exitCode=0 Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.240993 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lq9j" event={"ID":"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e","Type":"ContainerDied","Data":"8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61"} Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.241018 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lq9j" event={"ID":"0c08ee78-66ee-4c9a-b779-ab667bdb8e7e","Type":"ContainerDied","Data":"36dc29dc1502acefd7ae3ba88085b02883d8d1f4a20a1127096d7ec23d7ca3a4"} Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.241035 4944 scope.go:117] "RemoveContainer" containerID="8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61" Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.241167 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lq9j" Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.259823 4944 scope.go:117] "RemoveContainer" containerID="062fa447f1d00122f7151c702827e667fd88d7146e252beac4bbcd197b32bd03" Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.269326 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lq9j"] Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.274905 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lq9j"] Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.285510 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" path="/var/lib/kubelet/pods/0c08ee78-66ee-4c9a-b779-ab667bdb8e7e/volumes" Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.295349 4944 scope.go:117] "RemoveContainer" containerID="28ba6ef9c3cb23fcfbbbba19deb29f4033668d5b2720dd1b22322760b6619783" Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.317261 4944 scope.go:117] "RemoveContainer" containerID="8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61" Nov 24 10:10:30 crc kubenswrapper[4944]: E1124 10:10:30.317675 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61\": container with ID starting with 8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61 not found: ID does not exist" containerID="8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61" Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.317708 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61"} err="failed to get container status \"8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61\": rpc error: code = NotFound desc = could not find container \"8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61\": container with ID starting with 8f23d804d761644fa7a5120a3113cfbde75c04991b6128cd720bade98db44d61 not found: ID does not exist" Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.317729 4944 scope.go:117] "RemoveContainer" containerID="062fa447f1d00122f7151c702827e667fd88d7146e252beac4bbcd197b32bd03" Nov 24 10:10:30 crc kubenswrapper[4944]: E1124 10:10:30.319927 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"062fa447f1d00122f7151c702827e667fd88d7146e252beac4bbcd197b32bd03\": container with ID starting with 062fa447f1d00122f7151c702827e667fd88d7146e252beac4bbcd197b32bd03 not found: ID does not exist" containerID="062fa447f1d00122f7151c702827e667fd88d7146e252beac4bbcd197b32bd03" Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.319970 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"062fa447f1d00122f7151c702827e667fd88d7146e252beac4bbcd197b32bd03"} err="failed to get container status \"062fa447f1d00122f7151c702827e667fd88d7146e252beac4bbcd197b32bd03\": rpc error: code = NotFound desc = could not find container \"062fa447f1d00122f7151c702827e667fd88d7146e252beac4bbcd197b32bd03\": container with ID starting with 062fa447f1d00122f7151c702827e667fd88d7146e252beac4bbcd197b32bd03 not found: ID does not exist" Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.320001 4944 scope.go:117] "RemoveContainer" containerID="28ba6ef9c3cb23fcfbbbba19deb29f4033668d5b2720dd1b22322760b6619783" Nov 24 10:10:30 crc kubenswrapper[4944]: E1124 10:10:30.320322 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28ba6ef9c3cb23fcfbbbba19deb29f4033668d5b2720dd1b22322760b6619783\": container with ID starting with 28ba6ef9c3cb23fcfbbbba19deb29f4033668d5b2720dd1b22322760b6619783 not found: ID does not exist" containerID="28ba6ef9c3cb23fcfbbbba19deb29f4033668d5b2720dd1b22322760b6619783" Nov 24 10:10:30 crc kubenswrapper[4944]: I1124 10:10:30.320344 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28ba6ef9c3cb23fcfbbbba19deb29f4033668d5b2720dd1b22322760b6619783"} err="failed to get container status \"28ba6ef9c3cb23fcfbbbba19deb29f4033668d5b2720dd1b22322760b6619783\": rpc error: code = NotFound desc = could not find container \"28ba6ef9c3cb23fcfbbbba19deb29f4033668d5b2720dd1b22322760b6619783\": container with ID starting with 28ba6ef9c3cb23fcfbbbba19deb29f4033668d5b2720dd1b22322760b6619783 not found: ID does not exist" Nov 24 10:10:31 crc kubenswrapper[4944]: I1124 10:10:31.875162 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 10:10:31 crc kubenswrapper[4944]: E1124 10:10:31.877619 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" containerName="registry-server" Nov 24 10:10:31 crc kubenswrapper[4944]: I1124 10:10:31.877645 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" containerName="registry-server" Nov 24 10:10:31 crc kubenswrapper[4944]: E1124 10:10:31.877663 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" containerName="extract-content" Nov 24 10:10:31 crc kubenswrapper[4944]: I1124 10:10:31.877669 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" containerName="extract-content" Nov 24 10:10:31 crc kubenswrapper[4944]: E1124 10:10:31.877695 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79adaf85-ae0b-447e-a908-a78b796edc12" containerName="mariadb-client-4-default" Nov 24 10:10:31 crc kubenswrapper[4944]: I1124 10:10:31.877703 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="79adaf85-ae0b-447e-a908-a78b796edc12" containerName="mariadb-client-4-default" Nov 24 10:10:31 crc kubenswrapper[4944]: E1124 10:10:31.877723 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" containerName="extract-utilities" Nov 24 10:10:31 crc kubenswrapper[4944]: I1124 10:10:31.877731 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" containerName="extract-utilities" Nov 24 10:10:31 crc kubenswrapper[4944]: I1124 10:10:31.877876 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c08ee78-66ee-4c9a-b779-ab667bdb8e7e" containerName="registry-server" Nov 24 10:10:31 crc kubenswrapper[4944]: I1124 10:10:31.877905 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="79adaf85-ae0b-447e-a908-a78b796edc12" containerName="mariadb-client-4-default" Nov 24 10:10:31 crc kubenswrapper[4944]: I1124 10:10:31.879229 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 10:10:31 crc kubenswrapper[4944]: I1124 10:10:31.881230 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 10:10:31 crc kubenswrapper[4944]: I1124 10:10:31.882000 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-v4r8j" Nov 24 10:10:31 crc kubenswrapper[4944]: I1124 10:10:31.972809 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbpvv\" (UniqueName: \"kubernetes.io/projected/4cc0f310-cb10-4ebd-bcfc-ae74a721f06b-kube-api-access-wbpvv\") pod \"mariadb-client-5-default\" (UID: \"4cc0f310-cb10-4ebd-bcfc-ae74a721f06b\") " pod="openstack/mariadb-client-5-default" Nov 24 10:10:32 crc kubenswrapper[4944]: I1124 10:10:32.074751 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbpvv\" (UniqueName: \"kubernetes.io/projected/4cc0f310-cb10-4ebd-bcfc-ae74a721f06b-kube-api-access-wbpvv\") pod \"mariadb-client-5-default\" (UID: \"4cc0f310-cb10-4ebd-bcfc-ae74a721f06b\") " pod="openstack/mariadb-client-5-default" Nov 24 10:10:32 crc kubenswrapper[4944]: I1124 10:10:32.099849 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbpvv\" (UniqueName: \"kubernetes.io/projected/4cc0f310-cb10-4ebd-bcfc-ae74a721f06b-kube-api-access-wbpvv\") pod \"mariadb-client-5-default\" (UID: \"4cc0f310-cb10-4ebd-bcfc-ae74a721f06b\") " pod="openstack/mariadb-client-5-default" Nov 24 10:10:32 crc kubenswrapper[4944]: I1124 10:10:32.200099 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 10:10:32 crc kubenswrapper[4944]: I1124 10:10:32.726729 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 10:10:32 crc kubenswrapper[4944]: W1124 10:10:32.729127 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cc0f310_cb10_4ebd_bcfc_ae74a721f06b.slice/crio-32562f3ea413ee49f52295f5570191d1927aaf4e167361cec04f7a86b5aa55e8 WatchSource:0}: Error finding container 32562f3ea413ee49f52295f5570191d1927aaf4e167361cec04f7a86b5aa55e8: Status 404 returned error can't find the container with id 32562f3ea413ee49f52295f5570191d1927aaf4e167361cec04f7a86b5aa55e8 Nov 24 10:10:33 crc kubenswrapper[4944]: I1124 10:10:33.281316 4944 generic.go:334] "Generic (PLEG): container finished" podID="4cc0f310-cb10-4ebd-bcfc-ae74a721f06b" containerID="d6645eb0c62d5fd436446e5734ef76044e695c72011c8d71b17be3403ce8ec10" exitCode=0 Nov 24 10:10:33 crc kubenswrapper[4944]: I1124 10:10:33.281363 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"4cc0f310-cb10-4ebd-bcfc-ae74a721f06b","Type":"ContainerDied","Data":"d6645eb0c62d5fd436446e5734ef76044e695c72011c8d71b17be3403ce8ec10"} Nov 24 10:10:33 crc kubenswrapper[4944]: I1124 10:10:33.281393 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"4cc0f310-cb10-4ebd-bcfc-ae74a721f06b","Type":"ContainerStarted","Data":"32562f3ea413ee49f52295f5570191d1927aaf4e167361cec04f7a86b5aa55e8"} Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.314534 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-485r5"] Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.315842 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-485r5"] Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.315922 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.406499 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-catalog-content\") pod \"certified-operators-485r5\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.406857 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-utilities\") pod \"certified-operators-485r5\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.406878 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l9m8\" (UniqueName: \"kubernetes.io/projected/70a64252-8000-477c-a383-56c81e4c047e-kube-api-access-9l9m8\") pod \"certified-operators-485r5\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.511980 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-catalog-content\") pod \"certified-operators-485r5\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.512062 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-utilities\") pod \"certified-operators-485r5\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.512097 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l9m8\" (UniqueName: \"kubernetes.io/projected/70a64252-8000-477c-a383-56c81e4c047e-kube-api-access-9l9m8\") pod \"certified-operators-485r5\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.512688 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-utilities\") pod \"certified-operators-485r5\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.512686 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-catalog-content\") pod \"certified-operators-485r5\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.539133 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l9m8\" (UniqueName: \"kubernetes.io/projected/70a64252-8000-477c-a383-56c81e4c047e-kube-api-access-9l9m8\") pod \"certified-operators-485r5\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.646607 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.737254 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 10:10:34 crc kubenswrapper[4944]: I1124 10:10:34.766055 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_4cc0f310-cb10-4ebd-bcfc-ae74a721f06b/mariadb-client-5-default/0.log" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:34.794925 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:34.816924 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbpvv\" (UniqueName: \"kubernetes.io/projected/4cc0f310-cb10-4ebd-bcfc-ae74a721f06b-kube-api-access-wbpvv\") pod \"4cc0f310-cb10-4ebd-bcfc-ae74a721f06b\" (UID: \"4cc0f310-cb10-4ebd-bcfc-ae74a721f06b\") " Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:34.846335 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:34.863518 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc0f310-cb10-4ebd-bcfc-ae74a721f06b-kube-api-access-wbpvv" (OuterVolumeSpecName: "kube-api-access-wbpvv") pod "4cc0f310-cb10-4ebd-bcfc-ae74a721f06b" (UID: "4cc0f310-cb10-4ebd-bcfc-ae74a721f06b"). InnerVolumeSpecName "kube-api-access-wbpvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:34.919280 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbpvv\" (UniqueName: \"kubernetes.io/projected/4cc0f310-cb10-4ebd-bcfc-ae74a721f06b-kube-api-access-wbpvv\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:34.923780 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 10:10:35 crc kubenswrapper[4944]: E1124 10:10:34.924195 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc0f310-cb10-4ebd-bcfc-ae74a721f06b" containerName="mariadb-client-5-default" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:34.924210 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc0f310-cb10-4ebd-bcfc-ae74a721f06b" containerName="mariadb-client-5-default" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:34.924419 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc0f310-cb10-4ebd-bcfc-ae74a721f06b" containerName="mariadb-client-5-default" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:34.926004 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:34.935800 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:35.021115 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzg2c\" (UniqueName: \"kubernetes.io/projected/0010d296-b55a-4980-8d05-3c5ec2f97c39-kube-api-access-fzg2c\") pod \"mariadb-client-6-default\" (UID: \"0010d296-b55a-4980-8d05-3c5ec2f97c39\") " pod="openstack/mariadb-client-6-default" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:35.122870 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzg2c\" (UniqueName: \"kubernetes.io/projected/0010d296-b55a-4980-8d05-3c5ec2f97c39-kube-api-access-fzg2c\") pod \"mariadb-client-6-default\" (UID: \"0010d296-b55a-4980-8d05-3c5ec2f97c39\") " pod="openstack/mariadb-client-6-default" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:35.139590 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzg2c\" (UniqueName: \"kubernetes.io/projected/0010d296-b55a-4980-8d05-3c5ec2f97c39-kube-api-access-fzg2c\") pod \"mariadb-client-6-default\" (UID: \"0010d296-b55a-4980-8d05-3c5ec2f97c39\") " pod="openstack/mariadb-client-6-default" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:35.252364 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:35.303305 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32562f3ea413ee49f52295f5570191d1927aaf4e167361cec04f7a86b5aa55e8" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:35.303368 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:35.793561 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-485r5"] Nov 24 10:10:35 crc kubenswrapper[4944]: W1124 10:10:35.798149 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70a64252_8000_477c_a383_56c81e4c047e.slice/crio-c73eb821fcefeb177017e78219e57111280171e0d965aecdccdda989849db659 WatchSource:0}: Error finding container c73eb821fcefeb177017e78219e57111280171e0d965aecdccdda989849db659: Status 404 returned error can't find the container with id c73eb821fcefeb177017e78219e57111280171e0d965aecdccdda989849db659 Nov 24 10:10:35 crc kubenswrapper[4944]: W1124 10:10:35.914985 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0010d296_b55a_4980_8d05_3c5ec2f97c39.slice/crio-1cfbd9e4025a804003bb38bdb39b8ab8a5f170cb5407fb1d08b8f781d24b31d8 WatchSource:0}: Error finding container 1cfbd9e4025a804003bb38bdb39b8ab8a5f170cb5407fb1d08b8f781d24b31d8: Status 404 returned error can't find the container with id 1cfbd9e4025a804003bb38bdb39b8ab8a5f170cb5407fb1d08b8f781d24b31d8 Nov 24 10:10:35 crc kubenswrapper[4944]: I1124 10:10:35.915812 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 10:10:36 crc kubenswrapper[4944]: I1124 10:10:36.287251 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cc0f310-cb10-4ebd-bcfc-ae74a721f06b" path="/var/lib/kubelet/pods/4cc0f310-cb10-4ebd-bcfc-ae74a721f06b/volumes" Nov 24 10:10:36 crc kubenswrapper[4944]: I1124 10:10:36.311473 4944 generic.go:334] "Generic (PLEG): container finished" podID="70a64252-8000-477c-a383-56c81e4c047e" containerID="45a5de5ca66ad9736406ed78083d1e417b046ab84e48e0b4af1c2310a82da710" exitCode=0 Nov 24 10:10:36 crc kubenswrapper[4944]: I1124 10:10:36.311597 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-485r5" event={"ID":"70a64252-8000-477c-a383-56c81e4c047e","Type":"ContainerDied","Data":"45a5de5ca66ad9736406ed78083d1e417b046ab84e48e0b4af1c2310a82da710"} Nov 24 10:10:36 crc kubenswrapper[4944]: I1124 10:10:36.311647 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-485r5" event={"ID":"70a64252-8000-477c-a383-56c81e4c047e","Type":"ContainerStarted","Data":"c73eb821fcefeb177017e78219e57111280171e0d965aecdccdda989849db659"} Nov 24 10:10:36 crc kubenswrapper[4944]: I1124 10:10:36.314679 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"0010d296-b55a-4980-8d05-3c5ec2f97c39","Type":"ContainerStarted","Data":"1e9c7814f90a47b09017a40258d37a49b1de78b2c47ee2a620422ed80d0b6610"} Nov 24 10:10:36 crc kubenswrapper[4944]: I1124 10:10:36.314712 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"0010d296-b55a-4980-8d05-3c5ec2f97c39","Type":"ContainerStarted","Data":"1cfbd9e4025a804003bb38bdb39b8ab8a5f170cb5407fb1d08b8f781d24b31d8"} Nov 24 10:10:36 crc kubenswrapper[4944]: I1124 10:10:36.348680 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=2.348656245 podStartE2EDuration="2.348656245s" podCreationTimestamp="2025-11-24 10:10:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:10:36.343093427 +0000 UTC m=+4696.877533899" watchObservedRunningTime="2025-11-24 10:10:36.348656245 +0000 UTC m=+4696.883096707" Nov 24 10:10:37 crc kubenswrapper[4944]: I1124 10:10:37.322176 4944 generic.go:334] "Generic (PLEG): container finished" podID="0010d296-b55a-4980-8d05-3c5ec2f97c39" containerID="1e9c7814f90a47b09017a40258d37a49b1de78b2c47ee2a620422ed80d0b6610" exitCode=1 Nov 24 10:10:37 crc kubenswrapper[4944]: I1124 10:10:37.322247 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"0010d296-b55a-4980-8d05-3c5ec2f97c39","Type":"ContainerDied","Data":"1e9c7814f90a47b09017a40258d37a49b1de78b2c47ee2a620422ed80d0b6610"} Nov 24 10:10:37 crc kubenswrapper[4944]: I1124 10:10:37.325503 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-485r5" event={"ID":"70a64252-8000-477c-a383-56c81e4c047e","Type":"ContainerStarted","Data":"eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0"} Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.340322 4944 generic.go:334] "Generic (PLEG): container finished" podID="70a64252-8000-477c-a383-56c81e4c047e" containerID="eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0" exitCode=0 Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.340572 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-485r5" event={"ID":"70a64252-8000-477c-a383-56c81e4c047e","Type":"ContainerDied","Data":"eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0"} Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.706313 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.738454 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.743872 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.781190 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzg2c\" (UniqueName: \"kubernetes.io/projected/0010d296-b55a-4980-8d05-3c5ec2f97c39-kube-api-access-fzg2c\") pod \"0010d296-b55a-4980-8d05-3c5ec2f97c39\" (UID: \"0010d296-b55a-4980-8d05-3c5ec2f97c39\") " Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.792523 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0010d296-b55a-4980-8d05-3c5ec2f97c39-kube-api-access-fzg2c" (OuterVolumeSpecName: "kube-api-access-fzg2c") pod "0010d296-b55a-4980-8d05-3c5ec2f97c39" (UID: "0010d296-b55a-4980-8d05-3c5ec2f97c39"). InnerVolumeSpecName "kube-api-access-fzg2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.858634 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 10:10:38 crc kubenswrapper[4944]: E1124 10:10:38.859262 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0010d296-b55a-4980-8d05-3c5ec2f97c39" containerName="mariadb-client-6-default" Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.859468 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0010d296-b55a-4980-8d05-3c5ec2f97c39" containerName="mariadb-client-6-default" Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.863303 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0010d296-b55a-4980-8d05-3c5ec2f97c39" containerName="mariadb-client-6-default" Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.865400 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.870716 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.884098 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzg2c\" (UniqueName: \"kubernetes.io/projected/0010d296-b55a-4980-8d05-3c5ec2f97c39-kube-api-access-fzg2c\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:38 crc kubenswrapper[4944]: I1124 10:10:38.985423 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wglvx\" (UniqueName: \"kubernetes.io/projected/a63c794e-e8e1-42a1-a231-f54351003c22-kube-api-access-wglvx\") pod \"mariadb-client-7-default\" (UID: \"a63c794e-e8e1-42a1-a231-f54351003c22\") " pod="openstack/mariadb-client-7-default" Nov 24 10:10:39 crc kubenswrapper[4944]: I1124 10:10:39.089474 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wglvx\" (UniqueName: \"kubernetes.io/projected/a63c794e-e8e1-42a1-a231-f54351003c22-kube-api-access-wglvx\") pod \"mariadb-client-7-default\" (UID: \"a63c794e-e8e1-42a1-a231-f54351003c22\") " pod="openstack/mariadb-client-7-default" Nov 24 10:10:39 crc kubenswrapper[4944]: I1124 10:10:39.116650 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wglvx\" (UniqueName: \"kubernetes.io/projected/a63c794e-e8e1-42a1-a231-f54351003c22-kube-api-access-wglvx\") pod \"mariadb-client-7-default\" (UID: \"a63c794e-e8e1-42a1-a231-f54351003c22\") " pod="openstack/mariadb-client-7-default" Nov 24 10:10:39 crc kubenswrapper[4944]: I1124 10:10:39.198075 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 10:10:39 crc kubenswrapper[4944]: I1124 10:10:39.350471 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cfbd9e4025a804003bb38bdb39b8ab8a5f170cb5407fb1d08b8f781d24b31d8" Nov 24 10:10:39 crc kubenswrapper[4944]: I1124 10:10:39.350541 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 24 10:10:39 crc kubenswrapper[4944]: I1124 10:10:39.354424 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-485r5" event={"ID":"70a64252-8000-477c-a383-56c81e4c047e","Type":"ContainerStarted","Data":"2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3"} Nov 24 10:10:39 crc kubenswrapper[4944]: I1124 10:10:39.378114 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-485r5" podStartSLOduration=2.92438258 podStartE2EDuration="5.378090468s" podCreationTimestamp="2025-11-24 10:10:34 +0000 UTC" firstStartedPulling="2025-11-24 10:10:36.312975879 +0000 UTC m=+4696.847416341" lastFinishedPulling="2025-11-24 10:10:38.766683767 +0000 UTC m=+4699.301124229" observedRunningTime="2025-11-24 10:10:39.370136355 +0000 UTC m=+4699.904576837" watchObservedRunningTime="2025-11-24 10:10:39.378090468 +0000 UTC m=+4699.912530940" Nov 24 10:10:39 crc kubenswrapper[4944]: I1124 10:10:39.752005 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 10:10:39 crc kubenswrapper[4944]: W1124 10:10:39.756780 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda63c794e_e8e1_42a1_a231_f54351003c22.slice/crio-e18792e0f0532f9430addac4eb8a2b5fd82d249ed85a0a3a857e6532bb9d20de WatchSource:0}: Error finding container e18792e0f0532f9430addac4eb8a2b5fd82d249ed85a0a3a857e6532bb9d20de: Status 404 returned error can't find the container with id e18792e0f0532f9430addac4eb8a2b5fd82d249ed85a0a3a857e6532bb9d20de Nov 24 10:10:40 crc kubenswrapper[4944]: I1124 10:10:40.288180 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0010d296-b55a-4980-8d05-3c5ec2f97c39" path="/var/lib/kubelet/pods/0010d296-b55a-4980-8d05-3c5ec2f97c39/volumes" Nov 24 10:10:40 crc kubenswrapper[4944]: I1124 10:10:40.362425 4944 generic.go:334] "Generic (PLEG): container finished" podID="a63c794e-e8e1-42a1-a231-f54351003c22" containerID="8914a4dca5a6c4651a685a3b45bcd3139cfafc7e786ca1aa465535eb1ba36880" exitCode=0 Nov 24 10:10:40 crc kubenswrapper[4944]: I1124 10:10:40.362483 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"a63c794e-e8e1-42a1-a231-f54351003c22","Type":"ContainerDied","Data":"8914a4dca5a6c4651a685a3b45bcd3139cfafc7e786ca1aa465535eb1ba36880"} Nov 24 10:10:40 crc kubenswrapper[4944]: I1124 10:10:40.362532 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"a63c794e-e8e1-42a1-a231-f54351003c22","Type":"ContainerStarted","Data":"e18792e0f0532f9430addac4eb8a2b5fd82d249ed85a0a3a857e6532bb9d20de"} Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.715283 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.733703 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_a63c794e-e8e1-42a1-a231-f54351003c22/mariadb-client-7-default/0.log" Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.760706 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.764932 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.834115 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wglvx\" (UniqueName: \"kubernetes.io/projected/a63c794e-e8e1-42a1-a231-f54351003c22-kube-api-access-wglvx\") pod \"a63c794e-e8e1-42a1-a231-f54351003c22\" (UID: \"a63c794e-e8e1-42a1-a231-f54351003c22\") " Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.839003 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a63c794e-e8e1-42a1-a231-f54351003c22-kube-api-access-wglvx" (OuterVolumeSpecName: "kube-api-access-wglvx") pod "a63c794e-e8e1-42a1-a231-f54351003c22" (UID: "a63c794e-e8e1-42a1-a231-f54351003c22"). InnerVolumeSpecName "kube-api-access-wglvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.928796 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Nov 24 10:10:41 crc kubenswrapper[4944]: E1124 10:10:41.929171 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63c794e-e8e1-42a1-a231-f54351003c22" containerName="mariadb-client-7-default" Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.929185 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63c794e-e8e1-42a1-a231-f54351003c22" containerName="mariadb-client-7-default" Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.929328 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a63c794e-e8e1-42a1-a231-f54351003c22" containerName="mariadb-client-7-default" Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.929903 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.936435 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wglvx\" (UniqueName: \"kubernetes.io/projected/a63c794e-e8e1-42a1-a231-f54351003c22-kube-api-access-wglvx\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:41 crc kubenswrapper[4944]: I1124 10:10:41.946230 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 10:10:42 crc kubenswrapper[4944]: I1124 10:10:42.037683 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fxw8\" (UniqueName: \"kubernetes.io/projected/1b44bdae-f15d-4e09-bc26-f31497528be6-kube-api-access-9fxw8\") pod \"mariadb-client-2\" (UID: \"1b44bdae-f15d-4e09-bc26-f31497528be6\") " pod="openstack/mariadb-client-2" Nov 24 10:10:42 crc kubenswrapper[4944]: I1124 10:10:42.139993 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fxw8\" (UniqueName: \"kubernetes.io/projected/1b44bdae-f15d-4e09-bc26-f31497528be6-kube-api-access-9fxw8\") pod \"mariadb-client-2\" (UID: \"1b44bdae-f15d-4e09-bc26-f31497528be6\") " pod="openstack/mariadb-client-2" Nov 24 10:10:42 crc kubenswrapper[4944]: I1124 10:10:42.157601 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fxw8\" (UniqueName: \"kubernetes.io/projected/1b44bdae-f15d-4e09-bc26-f31497528be6-kube-api-access-9fxw8\") pod \"mariadb-client-2\" (UID: \"1b44bdae-f15d-4e09-bc26-f31497528be6\") " pod="openstack/mariadb-client-2" Nov 24 10:10:42 crc kubenswrapper[4944]: I1124 10:10:42.247683 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 10:10:42 crc kubenswrapper[4944]: I1124 10:10:42.284648 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a63c794e-e8e1-42a1-a231-f54351003c22" path="/var/lib/kubelet/pods/a63c794e-e8e1-42a1-a231-f54351003c22/volumes" Nov 24 10:10:42 crc kubenswrapper[4944]: I1124 10:10:42.377782 4944 scope.go:117] "RemoveContainer" containerID="8914a4dca5a6c4651a685a3b45bcd3139cfafc7e786ca1aa465535eb1ba36880" Nov 24 10:10:42 crc kubenswrapper[4944]: I1124 10:10:42.377848 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 24 10:10:42 crc kubenswrapper[4944]: I1124 10:10:42.753819 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 10:10:43 crc kubenswrapper[4944]: I1124 10:10:43.385819 4944 generic.go:334] "Generic (PLEG): container finished" podID="1b44bdae-f15d-4e09-bc26-f31497528be6" containerID="53e7aee10dd9a999650fb5910c0a7ec6104f282349fc6388a42c8d33f6c80f53" exitCode=0 Nov 24 10:10:43 crc kubenswrapper[4944]: I1124 10:10:43.385887 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"1b44bdae-f15d-4e09-bc26-f31497528be6","Type":"ContainerDied","Data":"53e7aee10dd9a999650fb5910c0a7ec6104f282349fc6388a42c8d33f6c80f53"} Nov 24 10:10:43 crc kubenswrapper[4944]: I1124 10:10:43.386178 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"1b44bdae-f15d-4e09-bc26-f31497528be6","Type":"ContainerStarted","Data":"805b77c9d2598806963f70767c99f5c8bb58cf5f03d853421abae2e40cd97e2b"} Nov 24 10:10:44 crc kubenswrapper[4944]: I1124 10:10:44.646890 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:44 crc kubenswrapper[4944]: I1124 10:10:44.648764 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:44 crc kubenswrapper[4944]: I1124 10:10:44.693685 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:44 crc kubenswrapper[4944]: I1124 10:10:44.728253 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 10:10:44 crc kubenswrapper[4944]: I1124 10:10:44.748490 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_1b44bdae-f15d-4e09-bc26-f31497528be6/mariadb-client-2/0.log" Nov 24 10:10:44 crc kubenswrapper[4944]: I1124 10:10:44.802615 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 10:10:44 crc kubenswrapper[4944]: I1124 10:10:44.810670 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Nov 24 10:10:44 crc kubenswrapper[4944]: I1124 10:10:44.882145 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fxw8\" (UniqueName: \"kubernetes.io/projected/1b44bdae-f15d-4e09-bc26-f31497528be6-kube-api-access-9fxw8\") pod \"1b44bdae-f15d-4e09-bc26-f31497528be6\" (UID: \"1b44bdae-f15d-4e09-bc26-f31497528be6\") " Nov 24 10:10:44 crc kubenswrapper[4944]: I1124 10:10:44.887587 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b44bdae-f15d-4e09-bc26-f31497528be6-kube-api-access-9fxw8" (OuterVolumeSpecName: "kube-api-access-9fxw8") pod "1b44bdae-f15d-4e09-bc26-f31497528be6" (UID: "1b44bdae-f15d-4e09-bc26-f31497528be6"). InnerVolumeSpecName "kube-api-access-9fxw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:10:44 crc kubenswrapper[4944]: I1124 10:10:44.984160 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fxw8\" (UniqueName: \"kubernetes.io/projected/1b44bdae-f15d-4e09-bc26-f31497528be6-kube-api-access-9fxw8\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:45 crc kubenswrapper[4944]: I1124 10:10:45.401220 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="805b77c9d2598806963f70767c99f5c8bb58cf5f03d853421abae2e40cd97e2b" Nov 24 10:10:45 crc kubenswrapper[4944]: I1124 10:10:45.401225 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 24 10:10:45 crc kubenswrapper[4944]: I1124 10:10:45.441199 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:46 crc kubenswrapper[4944]: I1124 10:10:46.288223 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b44bdae-f15d-4e09-bc26-f31497528be6" path="/var/lib/kubelet/pods/1b44bdae-f15d-4e09-bc26-f31497528be6/volumes" Nov 24 10:10:47 crc kubenswrapper[4944]: I1124 10:10:47.872118 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-485r5"] Nov 24 10:10:47 crc kubenswrapper[4944]: I1124 10:10:47.872621 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-485r5" podUID="70a64252-8000-477c-a383-56c81e4c047e" containerName="registry-server" containerID="cri-o://2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3" gracePeriod=2 Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.280517 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.421615 4944 generic.go:334] "Generic (PLEG): container finished" podID="70a64252-8000-477c-a383-56c81e4c047e" containerID="2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3" exitCode=0 Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.421651 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-485r5" event={"ID":"70a64252-8000-477c-a383-56c81e4c047e","Type":"ContainerDied","Data":"2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3"} Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.421676 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-485r5" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.421689 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-485r5" event={"ID":"70a64252-8000-477c-a383-56c81e4c047e","Type":"ContainerDied","Data":"c73eb821fcefeb177017e78219e57111280171e0d965aecdccdda989849db659"} Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.421708 4944 scope.go:117] "RemoveContainer" containerID="2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.435334 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-utilities\") pod \"70a64252-8000-477c-a383-56c81e4c047e\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.435373 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-catalog-content\") pod \"70a64252-8000-477c-a383-56c81e4c047e\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.435419 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l9m8\" (UniqueName: \"kubernetes.io/projected/70a64252-8000-477c-a383-56c81e4c047e-kube-api-access-9l9m8\") pod \"70a64252-8000-477c-a383-56c81e4c047e\" (UID: \"70a64252-8000-477c-a383-56c81e4c047e\") " Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.436444 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-utilities" (OuterVolumeSpecName: "utilities") pod "70a64252-8000-477c-a383-56c81e4c047e" (UID: "70a64252-8000-477c-a383-56c81e4c047e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.449560 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70a64252-8000-477c-a383-56c81e4c047e-kube-api-access-9l9m8" (OuterVolumeSpecName: "kube-api-access-9l9m8") pod "70a64252-8000-477c-a383-56c81e4c047e" (UID: "70a64252-8000-477c-a383-56c81e4c047e"). InnerVolumeSpecName "kube-api-access-9l9m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.449641 4944 scope.go:117] "RemoveContainer" containerID="eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.487628 4944 scope.go:117] "RemoveContainer" containerID="45a5de5ca66ad9736406ed78083d1e417b046ab84e48e0b4af1c2310a82da710" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.488578 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70a64252-8000-477c-a383-56c81e4c047e" (UID: "70a64252-8000-477c-a383-56c81e4c047e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.519763 4944 scope.go:117] "RemoveContainer" containerID="2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3" Nov 24 10:10:48 crc kubenswrapper[4944]: E1124 10:10:48.520569 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3\": container with ID starting with 2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3 not found: ID does not exist" containerID="2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.520605 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3"} err="failed to get container status \"2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3\": rpc error: code = NotFound desc = could not find container \"2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3\": container with ID starting with 2208aef27003506b48578c5e9bbb7f31acfd22aa6b54588ac7ccfa66d604c3e3 not found: ID does not exist" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.520628 4944 scope.go:117] "RemoveContainer" containerID="eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0" Nov 24 10:10:48 crc kubenswrapper[4944]: E1124 10:10:48.520910 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0\": container with ID starting with eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0 not found: ID does not exist" containerID="eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.520941 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0"} err="failed to get container status \"eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0\": rpc error: code = NotFound desc = could not find container \"eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0\": container with ID starting with eaef02a9d9b6ce86f3c749e01265f533a46e0ee6e2892c55cf4ad707ed7d93f0 not found: ID does not exist" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.520957 4944 scope.go:117] "RemoveContainer" containerID="45a5de5ca66ad9736406ed78083d1e417b046ab84e48e0b4af1c2310a82da710" Nov 24 10:10:48 crc kubenswrapper[4944]: E1124 10:10:48.521217 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45a5de5ca66ad9736406ed78083d1e417b046ab84e48e0b4af1c2310a82da710\": container with ID starting with 45a5de5ca66ad9736406ed78083d1e417b046ab84e48e0b4af1c2310a82da710 not found: ID does not exist" containerID="45a5de5ca66ad9736406ed78083d1e417b046ab84e48e0b4af1c2310a82da710" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.521238 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45a5de5ca66ad9736406ed78083d1e417b046ab84e48e0b4af1c2310a82da710"} err="failed to get container status \"45a5de5ca66ad9736406ed78083d1e417b046ab84e48e0b4af1c2310a82da710\": rpc error: code = NotFound desc = could not find container \"45a5de5ca66ad9736406ed78083d1e417b046ab84e48e0b4af1c2310a82da710\": container with ID starting with 45a5de5ca66ad9736406ed78083d1e417b046ab84e48e0b4af1c2310a82da710 not found: ID does not exist" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.537344 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.537377 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70a64252-8000-477c-a383-56c81e4c047e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.537387 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l9m8\" (UniqueName: \"kubernetes.io/projected/70a64252-8000-477c-a383-56c81e4c047e-kube-api-access-9l9m8\") on node \"crc\" DevicePath \"\"" Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.759945 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-485r5"] Nov 24 10:10:48 crc kubenswrapper[4944]: I1124 10:10:48.767535 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-485r5"] Nov 24 10:10:50 crc kubenswrapper[4944]: I1124 10:10:50.284522 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70a64252-8000-477c-a383-56c81e4c047e" path="/var/lib/kubelet/pods/70a64252-8000-477c-a383-56c81e4c047e/volumes" Nov 24 10:11:23 crc kubenswrapper[4944]: I1124 10:11:23.548684 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:11:23 crc kubenswrapper[4944]: I1124 10:11:23.549206 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:11:29 crc kubenswrapper[4944]: I1124 10:11:29.999552 4944 scope.go:117] "RemoveContainer" containerID="1face7647d6c31173d7d8661432523c4108f050516f757d27a0c690b660b3c8c" Nov 24 10:11:53 crc kubenswrapper[4944]: I1124 10:11:53.548858 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:11:53 crc kubenswrapper[4944]: I1124 10:11:53.549500 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:12:23 crc kubenswrapper[4944]: I1124 10:12:23.548816 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:12:23 crc kubenswrapper[4944]: I1124 10:12:23.549453 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:12:23 crc kubenswrapper[4944]: I1124 10:12:23.549525 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 10:12:23 crc kubenswrapper[4944]: I1124 10:12:23.550298 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"54a6b243aeb67222b25b473701ee151808da9cae0e39b5478f62186939a082c4"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 10:12:23 crc kubenswrapper[4944]: I1124 10:12:23.550371 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://54a6b243aeb67222b25b473701ee151808da9cae0e39b5478f62186939a082c4" gracePeriod=600 Nov 24 10:12:24 crc kubenswrapper[4944]: I1124 10:12:24.133547 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="54a6b243aeb67222b25b473701ee151808da9cae0e39b5478f62186939a082c4" exitCode=0 Nov 24 10:12:24 crc kubenswrapper[4944]: I1124 10:12:24.133616 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"54a6b243aeb67222b25b473701ee151808da9cae0e39b5478f62186939a082c4"} Nov 24 10:12:24 crc kubenswrapper[4944]: I1124 10:12:24.133892 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf"} Nov 24 10:12:24 crc kubenswrapper[4944]: I1124 10:12:24.133911 4944 scope.go:117] "RemoveContainer" containerID="c447978ca8a74a467d70372c48180511466837964921e0e6eab6046f1f30b641" Nov 24 10:14:23 crc kubenswrapper[4944]: I1124 10:14:23.548918 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:14:23 crc kubenswrapper[4944]: I1124 10:14:23.549590 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:14:53 crc kubenswrapper[4944]: I1124 10:14:53.549091 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:14:53 crc kubenswrapper[4944]: I1124 10:14:53.549624 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.147376 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr"] Nov 24 10:15:00 crc kubenswrapper[4944]: E1124 10:15:00.148816 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b44bdae-f15d-4e09-bc26-f31497528be6" containerName="mariadb-client-2" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.148840 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b44bdae-f15d-4e09-bc26-f31497528be6" containerName="mariadb-client-2" Nov 24 10:15:00 crc kubenswrapper[4944]: E1124 10:15:00.148869 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a64252-8000-477c-a383-56c81e4c047e" containerName="extract-content" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.148876 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a64252-8000-477c-a383-56c81e4c047e" containerName="extract-content" Nov 24 10:15:00 crc kubenswrapper[4944]: E1124 10:15:00.148893 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a64252-8000-477c-a383-56c81e4c047e" containerName="registry-server" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.148901 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a64252-8000-477c-a383-56c81e4c047e" containerName="registry-server" Nov 24 10:15:00 crc kubenswrapper[4944]: E1124 10:15:00.148914 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70a64252-8000-477c-a383-56c81e4c047e" containerName="extract-utilities" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.148920 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="70a64252-8000-477c-a383-56c81e4c047e" containerName="extract-utilities" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.149141 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="70a64252-8000-477c-a383-56c81e4c047e" containerName="registry-server" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.149311 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b44bdae-f15d-4e09-bc26-f31497528be6" containerName="mariadb-client-2" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.150226 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.152401 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.155137 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.170861 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr"] Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.277909 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjtcq\" (UniqueName: \"kubernetes.io/projected/47ead2de-7b54-48d8-a5ec-f292678c1966-kube-api-access-wjtcq\") pod \"collect-profiles-29399655-xplzr\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.278027 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47ead2de-7b54-48d8-a5ec-f292678c1966-config-volume\") pod \"collect-profiles-29399655-xplzr\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.278080 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47ead2de-7b54-48d8-a5ec-f292678c1966-secret-volume\") pod \"collect-profiles-29399655-xplzr\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.379790 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjtcq\" (UniqueName: \"kubernetes.io/projected/47ead2de-7b54-48d8-a5ec-f292678c1966-kube-api-access-wjtcq\") pod \"collect-profiles-29399655-xplzr\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.379869 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47ead2de-7b54-48d8-a5ec-f292678c1966-config-volume\") pod \"collect-profiles-29399655-xplzr\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.379888 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47ead2de-7b54-48d8-a5ec-f292678c1966-secret-volume\") pod \"collect-profiles-29399655-xplzr\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.381065 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47ead2de-7b54-48d8-a5ec-f292678c1966-config-volume\") pod \"collect-profiles-29399655-xplzr\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.385713 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47ead2de-7b54-48d8-a5ec-f292678c1966-secret-volume\") pod \"collect-profiles-29399655-xplzr\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.394885 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjtcq\" (UniqueName: \"kubernetes.io/projected/47ead2de-7b54-48d8-a5ec-f292678c1966-kube-api-access-wjtcq\") pod \"collect-profiles-29399655-xplzr\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.471565 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:00 crc kubenswrapper[4944]: I1124 10:15:00.899468 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr"] Nov 24 10:15:01 crc kubenswrapper[4944]: I1124 10:15:01.407855 4944 generic.go:334] "Generic (PLEG): container finished" podID="47ead2de-7b54-48d8-a5ec-f292678c1966" containerID="a2458a1342744ce198f6787bedc601a87b08fcfa6bccf41b35d86c2aa2a442a8" exitCode=0 Nov 24 10:15:01 crc kubenswrapper[4944]: I1124 10:15:01.408022 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" event={"ID":"47ead2de-7b54-48d8-a5ec-f292678c1966","Type":"ContainerDied","Data":"a2458a1342744ce198f6787bedc601a87b08fcfa6bccf41b35d86c2aa2a442a8"} Nov 24 10:15:01 crc kubenswrapper[4944]: I1124 10:15:01.408156 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" event={"ID":"47ead2de-7b54-48d8-a5ec-f292678c1966","Type":"ContainerStarted","Data":"8734af6821ac606492ba2e3f44c4b4ba5e92a47e504b01a4048ae345845856b1"} Nov 24 10:15:02 crc kubenswrapper[4944]: I1124 10:15:02.706289 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:02 crc kubenswrapper[4944]: I1124 10:15:02.734514 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47ead2de-7b54-48d8-a5ec-f292678c1966-config-volume\") pod \"47ead2de-7b54-48d8-a5ec-f292678c1966\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " Nov 24 10:15:02 crc kubenswrapper[4944]: I1124 10:15:02.734679 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjtcq\" (UniqueName: \"kubernetes.io/projected/47ead2de-7b54-48d8-a5ec-f292678c1966-kube-api-access-wjtcq\") pod \"47ead2de-7b54-48d8-a5ec-f292678c1966\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " Nov 24 10:15:02 crc kubenswrapper[4944]: I1124 10:15:02.734736 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47ead2de-7b54-48d8-a5ec-f292678c1966-secret-volume\") pod \"47ead2de-7b54-48d8-a5ec-f292678c1966\" (UID: \"47ead2de-7b54-48d8-a5ec-f292678c1966\") " Nov 24 10:15:02 crc kubenswrapper[4944]: I1124 10:15:02.740612 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ead2de-7b54-48d8-a5ec-f292678c1966-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "47ead2de-7b54-48d8-a5ec-f292678c1966" (UID: "47ead2de-7b54-48d8-a5ec-f292678c1966"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:15:02 crc kubenswrapper[4944]: I1124 10:15:02.745165 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47ead2de-7b54-48d8-a5ec-f292678c1966-config-volume" (OuterVolumeSpecName: "config-volume") pod "47ead2de-7b54-48d8-a5ec-f292678c1966" (UID: "47ead2de-7b54-48d8-a5ec-f292678c1966"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:15:02 crc kubenswrapper[4944]: I1124 10:15:02.745268 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47ead2de-7b54-48d8-a5ec-f292678c1966-kube-api-access-wjtcq" (OuterVolumeSpecName: "kube-api-access-wjtcq") pod "47ead2de-7b54-48d8-a5ec-f292678c1966" (UID: "47ead2de-7b54-48d8-a5ec-f292678c1966"). InnerVolumeSpecName "kube-api-access-wjtcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:15:02 crc kubenswrapper[4944]: I1124 10:15:02.836252 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjtcq\" (UniqueName: \"kubernetes.io/projected/47ead2de-7b54-48d8-a5ec-f292678c1966-kube-api-access-wjtcq\") on node \"crc\" DevicePath \"\"" Nov 24 10:15:02 crc kubenswrapper[4944]: I1124 10:15:02.836307 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47ead2de-7b54-48d8-a5ec-f292678c1966-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 10:15:02 crc kubenswrapper[4944]: I1124 10:15:02.836320 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47ead2de-7b54-48d8-a5ec-f292678c1966-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 10:15:03 crc kubenswrapper[4944]: I1124 10:15:03.423246 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" event={"ID":"47ead2de-7b54-48d8-a5ec-f292678c1966","Type":"ContainerDied","Data":"8734af6821ac606492ba2e3f44c4b4ba5e92a47e504b01a4048ae345845856b1"} Nov 24 10:15:03 crc kubenswrapper[4944]: I1124 10:15:03.423297 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8734af6821ac606492ba2e3f44c4b4ba5e92a47e504b01a4048ae345845856b1" Nov 24 10:15:03 crc kubenswrapper[4944]: I1124 10:15:03.423317 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr" Nov 24 10:15:03 crc kubenswrapper[4944]: I1124 10:15:03.767665 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz"] Nov 24 10:15:03 crc kubenswrapper[4944]: I1124 10:15:03.773632 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399610-twmfz"] Nov 24 10:15:04 crc kubenswrapper[4944]: I1124 10:15:04.286320 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f5f63a4-f6aa-4656-83ec-884f6c4f52bf" path="/var/lib/kubelet/pods/3f5f63a4-f6aa-4656-83ec-884f6c4f52bf/volumes" Nov 24 10:15:23 crc kubenswrapper[4944]: I1124 10:15:23.548421 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:15:23 crc kubenswrapper[4944]: I1124 10:15:23.548988 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:15:23 crc kubenswrapper[4944]: I1124 10:15:23.549036 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 10:15:23 crc kubenswrapper[4944]: I1124 10:15:23.549606 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 10:15:23 crc kubenswrapper[4944]: I1124 10:15:23.549650 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" gracePeriod=600 Nov 24 10:15:24 crc kubenswrapper[4944]: E1124 10:15:24.202898 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:15:24 crc kubenswrapper[4944]: I1124 10:15:24.562035 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" exitCode=0 Nov 24 10:15:24 crc kubenswrapper[4944]: I1124 10:15:24.562104 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf"} Nov 24 10:15:24 crc kubenswrapper[4944]: I1124 10:15:24.562141 4944 scope.go:117] "RemoveContainer" containerID="54a6b243aeb67222b25b473701ee151808da9cae0e39b5478f62186939a082c4" Nov 24 10:15:24 crc kubenswrapper[4944]: I1124 10:15:24.562747 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:15:24 crc kubenswrapper[4944]: E1124 10:15:24.563011 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:15:30 crc kubenswrapper[4944]: I1124 10:15:30.135353 4944 scope.go:117] "RemoveContainer" containerID="91035c422b3f43b0d85fe46b85fdbaaea59ef42de502b69aef1d1ef5d9d6eacd" Nov 24 10:15:31 crc kubenswrapper[4944]: I1124 10:15:31.828716 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 10:15:31 crc kubenswrapper[4944]: E1124 10:15:31.829407 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ead2de-7b54-48d8-a5ec-f292678c1966" containerName="collect-profiles" Nov 24 10:15:31 crc kubenswrapper[4944]: I1124 10:15:31.829420 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ead2de-7b54-48d8-a5ec-f292678c1966" containerName="collect-profiles" Nov 24 10:15:31 crc kubenswrapper[4944]: I1124 10:15:31.829573 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ead2de-7b54-48d8-a5ec-f292678c1966" containerName="collect-profiles" Nov 24 10:15:31 crc kubenswrapper[4944]: I1124 10:15:31.830859 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 10:15:31 crc kubenswrapper[4944]: I1124 10:15:31.832927 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-v4r8j" Nov 24 10:15:31 crc kubenswrapper[4944]: I1124 10:15:31.839161 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 10:15:31 crc kubenswrapper[4944]: I1124 10:15:31.999839 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckw8g\" (UniqueName: \"kubernetes.io/projected/9ff0f1b1-a485-45a1-8a01-e59eb360445d-kube-api-access-ckw8g\") pod \"mariadb-copy-data\" (UID: \"9ff0f1b1-a485-45a1-8a01-e59eb360445d\") " pod="openstack/mariadb-copy-data" Nov 24 10:15:31 crc kubenswrapper[4944]: I1124 10:15:31.999907 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-258661ee-6084-4970-8b6a-a82e69415274\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-258661ee-6084-4970-8b6a-a82e69415274\") pod \"mariadb-copy-data\" (UID: \"9ff0f1b1-a485-45a1-8a01-e59eb360445d\") " pod="openstack/mariadb-copy-data" Nov 24 10:15:32 crc kubenswrapper[4944]: I1124 10:15:32.101412 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckw8g\" (UniqueName: \"kubernetes.io/projected/9ff0f1b1-a485-45a1-8a01-e59eb360445d-kube-api-access-ckw8g\") pod \"mariadb-copy-data\" (UID: \"9ff0f1b1-a485-45a1-8a01-e59eb360445d\") " pod="openstack/mariadb-copy-data" Nov 24 10:15:32 crc kubenswrapper[4944]: I1124 10:15:32.101479 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-258661ee-6084-4970-8b6a-a82e69415274\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-258661ee-6084-4970-8b6a-a82e69415274\") pod \"mariadb-copy-data\" (UID: \"9ff0f1b1-a485-45a1-8a01-e59eb360445d\") " pod="openstack/mariadb-copy-data" Nov 24 10:15:32 crc kubenswrapper[4944]: I1124 10:15:32.104284 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:15:32 crc kubenswrapper[4944]: I1124 10:15:32.104334 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-258661ee-6084-4970-8b6a-a82e69415274\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-258661ee-6084-4970-8b6a-a82e69415274\") pod \"mariadb-copy-data\" (UID: \"9ff0f1b1-a485-45a1-8a01-e59eb360445d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/27ebad340bdb7f7536c263f19a14a3192d77f0549c5490652a47e3b297c14569/globalmount\"" pod="openstack/mariadb-copy-data" Nov 24 10:15:32 crc kubenswrapper[4944]: I1124 10:15:32.123259 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckw8g\" (UniqueName: \"kubernetes.io/projected/9ff0f1b1-a485-45a1-8a01-e59eb360445d-kube-api-access-ckw8g\") pod \"mariadb-copy-data\" (UID: \"9ff0f1b1-a485-45a1-8a01-e59eb360445d\") " pod="openstack/mariadb-copy-data" Nov 24 10:15:32 crc kubenswrapper[4944]: I1124 10:15:32.132171 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-258661ee-6084-4970-8b6a-a82e69415274\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-258661ee-6084-4970-8b6a-a82e69415274\") pod \"mariadb-copy-data\" (UID: \"9ff0f1b1-a485-45a1-8a01-e59eb360445d\") " pod="openstack/mariadb-copy-data" Nov 24 10:15:32 crc kubenswrapper[4944]: I1124 10:15:32.178647 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 10:15:32 crc kubenswrapper[4944]: I1124 10:15:32.650165 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 10:15:33 crc kubenswrapper[4944]: I1124 10:15:33.664681 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"9ff0f1b1-a485-45a1-8a01-e59eb360445d","Type":"ContainerStarted","Data":"4c97f4fa04a0fb4a629e12fd1304cd09082f79eaebd148be0cdabcea44912e02"} Nov 24 10:15:33 crc kubenswrapper[4944]: I1124 10:15:33.664744 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"9ff0f1b1-a485-45a1-8a01-e59eb360445d","Type":"ContainerStarted","Data":"93f2df2de8ef452b3506018300d9e73624494ec2ac2fdcf9a6c3655efc913020"} Nov 24 10:15:33 crc kubenswrapper[4944]: I1124 10:15:33.680223 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.680206877 podStartE2EDuration="3.680206877s" podCreationTimestamp="2025-11-24 10:15:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:15:33.675158346 +0000 UTC m=+4994.209598808" watchObservedRunningTime="2025-11-24 10:15:33.680206877 +0000 UTC m=+4994.214647339" Nov 24 10:15:36 crc kubenswrapper[4944]: I1124 10:15:36.364829 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Nov 24 10:15:36 crc kubenswrapper[4944]: I1124 10:15:36.366703 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 10:15:36 crc kubenswrapper[4944]: I1124 10:15:36.375695 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 10:15:36 crc kubenswrapper[4944]: I1124 10:15:36.469378 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgtfh\" (UniqueName: \"kubernetes.io/projected/98ea80ff-c623-4f4a-8168-a8fa9e61543b-kube-api-access-tgtfh\") pod \"mariadb-client\" (UID: \"98ea80ff-c623-4f4a-8168-a8fa9e61543b\") " pod="openstack/mariadb-client" Nov 24 10:15:36 crc kubenswrapper[4944]: I1124 10:15:36.571222 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgtfh\" (UniqueName: \"kubernetes.io/projected/98ea80ff-c623-4f4a-8168-a8fa9e61543b-kube-api-access-tgtfh\") pod \"mariadb-client\" (UID: \"98ea80ff-c623-4f4a-8168-a8fa9e61543b\") " pod="openstack/mariadb-client" Nov 24 10:15:36 crc kubenswrapper[4944]: I1124 10:15:36.589956 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgtfh\" (UniqueName: \"kubernetes.io/projected/98ea80ff-c623-4f4a-8168-a8fa9e61543b-kube-api-access-tgtfh\") pod \"mariadb-client\" (UID: \"98ea80ff-c623-4f4a-8168-a8fa9e61543b\") " pod="openstack/mariadb-client" Nov 24 10:15:36 crc kubenswrapper[4944]: I1124 10:15:36.685746 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 10:15:37 crc kubenswrapper[4944]: I1124 10:15:37.184103 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 10:15:37 crc kubenswrapper[4944]: I1124 10:15:37.701199 4944 generic.go:334] "Generic (PLEG): container finished" podID="98ea80ff-c623-4f4a-8168-a8fa9e61543b" containerID="e835c2e61896806dff8001fa9a9ff6005548fda86b44d18d38904afa1e7e5107" exitCode=0 Nov 24 10:15:37 crc kubenswrapper[4944]: I1124 10:15:37.701267 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"98ea80ff-c623-4f4a-8168-a8fa9e61543b","Type":"ContainerDied","Data":"e835c2e61896806dff8001fa9a9ff6005548fda86b44d18d38904afa1e7e5107"} Nov 24 10:15:37 crc kubenswrapper[4944]: I1124 10:15:37.701305 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"98ea80ff-c623-4f4a-8168-a8fa9e61543b","Type":"ContainerStarted","Data":"724bbddbc47c60eb5bcf0f21b2e8255d6b9636597ef7e3156978156f4488f09b"} Nov 24 10:15:38 crc kubenswrapper[4944]: I1124 10:15:38.971865 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.042827 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_98ea80ff-c623-4f4a-8168-a8fa9e61543b/mariadb-client/0.log" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.067690 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.075254 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.136450 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgtfh\" (UniqueName: \"kubernetes.io/projected/98ea80ff-c623-4f4a-8168-a8fa9e61543b-kube-api-access-tgtfh\") pod \"98ea80ff-c623-4f4a-8168-a8fa9e61543b\" (UID: \"98ea80ff-c623-4f4a-8168-a8fa9e61543b\") " Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.141321 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98ea80ff-c623-4f4a-8168-a8fa9e61543b-kube-api-access-tgtfh" (OuterVolumeSpecName: "kube-api-access-tgtfh") pod "98ea80ff-c623-4f4a-8168-a8fa9e61543b" (UID: "98ea80ff-c623-4f4a-8168-a8fa9e61543b"). InnerVolumeSpecName "kube-api-access-tgtfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.187932 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Nov 24 10:15:39 crc kubenswrapper[4944]: E1124 10:15:39.188553 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ea80ff-c623-4f4a-8168-a8fa9e61543b" containerName="mariadb-client" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.188573 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ea80ff-c623-4f4a-8168-a8fa9e61543b" containerName="mariadb-client" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.188739 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="98ea80ff-c623-4f4a-8168-a8fa9e61543b" containerName="mariadb-client" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.189243 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.198566 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.237709 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m45bh\" (UniqueName: \"kubernetes.io/projected/dc26f34c-b7b7-4109-aef0-2792d712db47-kube-api-access-m45bh\") pod \"mariadb-client\" (UID: \"dc26f34c-b7b7-4109-aef0-2792d712db47\") " pod="openstack/mariadb-client" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.237865 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgtfh\" (UniqueName: \"kubernetes.io/projected/98ea80ff-c623-4f4a-8168-a8fa9e61543b-kube-api-access-tgtfh\") on node \"crc\" DevicePath \"\"" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.276692 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:15:39 crc kubenswrapper[4944]: E1124 10:15:39.276870 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.338994 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m45bh\" (UniqueName: \"kubernetes.io/projected/dc26f34c-b7b7-4109-aef0-2792d712db47-kube-api-access-m45bh\") pod \"mariadb-client\" (UID: \"dc26f34c-b7b7-4109-aef0-2792d712db47\") " pod="openstack/mariadb-client" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.354579 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m45bh\" (UniqueName: \"kubernetes.io/projected/dc26f34c-b7b7-4109-aef0-2792d712db47-kube-api-access-m45bh\") pod \"mariadb-client\" (UID: \"dc26f34c-b7b7-4109-aef0-2792d712db47\") " pod="openstack/mariadb-client" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.511452 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.716303 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="724bbddbc47c60eb5bcf0f21b2e8255d6b9636597ef7e3156978156f4488f09b" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.716353 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.733434 4944 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="98ea80ff-c623-4f4a-8168-a8fa9e61543b" podUID="dc26f34c-b7b7-4109-aef0-2792d712db47" Nov 24 10:15:39 crc kubenswrapper[4944]: I1124 10:15:39.918242 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 24 10:15:39 crc kubenswrapper[4944]: W1124 10:15:39.930439 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc26f34c_b7b7_4109_aef0_2792d712db47.slice/crio-640de1f1ad2fd2d04488f224e9793761d3df0d804c05c4eee1e72464cfab36ee WatchSource:0}: Error finding container 640de1f1ad2fd2d04488f224e9793761d3df0d804c05c4eee1e72464cfab36ee: Status 404 returned error can't find the container with id 640de1f1ad2fd2d04488f224e9793761d3df0d804c05c4eee1e72464cfab36ee Nov 24 10:15:40 crc kubenswrapper[4944]: I1124 10:15:40.287125 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98ea80ff-c623-4f4a-8168-a8fa9e61543b" path="/var/lib/kubelet/pods/98ea80ff-c623-4f4a-8168-a8fa9e61543b/volumes" Nov 24 10:15:40 crc kubenswrapper[4944]: I1124 10:15:40.725824 4944 generic.go:334] "Generic (PLEG): container finished" podID="dc26f34c-b7b7-4109-aef0-2792d712db47" containerID="06efd00b7795c0514a2b6678e84f3c8b8eb0b355e3c5d4cfdf8a1231790e0314" exitCode=0 Nov 24 10:15:40 crc kubenswrapper[4944]: I1124 10:15:40.725868 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"dc26f34c-b7b7-4109-aef0-2792d712db47","Type":"ContainerDied","Data":"06efd00b7795c0514a2b6678e84f3c8b8eb0b355e3c5d4cfdf8a1231790e0314"} Nov 24 10:15:40 crc kubenswrapper[4944]: I1124 10:15:40.725903 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"dc26f34c-b7b7-4109-aef0-2792d712db47","Type":"ContainerStarted","Data":"640de1f1ad2fd2d04488f224e9793761d3df0d804c05c4eee1e72464cfab36ee"} Nov 24 10:15:42 crc kubenswrapper[4944]: I1124 10:15:42.045837 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 10:15:42 crc kubenswrapper[4944]: I1124 10:15:42.064677 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_dc26f34c-b7b7-4109-aef0-2792d712db47/mariadb-client/0.log" Nov 24 10:15:42 crc kubenswrapper[4944]: I1124 10:15:42.087256 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m45bh\" (UniqueName: \"kubernetes.io/projected/dc26f34c-b7b7-4109-aef0-2792d712db47-kube-api-access-m45bh\") pod \"dc26f34c-b7b7-4109-aef0-2792d712db47\" (UID: \"dc26f34c-b7b7-4109-aef0-2792d712db47\") " Nov 24 10:15:42 crc kubenswrapper[4944]: I1124 10:15:42.090687 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Nov 24 10:15:42 crc kubenswrapper[4944]: I1124 10:15:42.097437 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc26f34c-b7b7-4109-aef0-2792d712db47-kube-api-access-m45bh" (OuterVolumeSpecName: "kube-api-access-m45bh") pod "dc26f34c-b7b7-4109-aef0-2792d712db47" (UID: "dc26f34c-b7b7-4109-aef0-2792d712db47"). InnerVolumeSpecName "kube-api-access-m45bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:15:42 crc kubenswrapper[4944]: I1124 10:15:42.097696 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Nov 24 10:15:42 crc kubenswrapper[4944]: I1124 10:15:42.189035 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m45bh\" (UniqueName: \"kubernetes.io/projected/dc26f34c-b7b7-4109-aef0-2792d712db47-kube-api-access-m45bh\") on node \"crc\" DevicePath \"\"" Nov 24 10:15:42 crc kubenswrapper[4944]: I1124 10:15:42.285400 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc26f34c-b7b7-4109-aef0-2792d712db47" path="/var/lib/kubelet/pods/dc26f34c-b7b7-4109-aef0-2792d712db47/volumes" Nov 24 10:15:42 crc kubenswrapper[4944]: I1124 10:15:42.740935 4944 scope.go:117] "RemoveContainer" containerID="06efd00b7795c0514a2b6678e84f3c8b8eb0b355e3c5d4cfdf8a1231790e0314" Nov 24 10:15:42 crc kubenswrapper[4944]: I1124 10:15:42.740978 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 24 10:15:52 crc kubenswrapper[4944]: I1124 10:15:52.276974 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:15:52 crc kubenswrapper[4944]: E1124 10:15:52.278589 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:16:04 crc kubenswrapper[4944]: I1124 10:16:04.276618 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:16:04 crc kubenswrapper[4944]: E1124 10:16:04.277346 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.727656 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 10:16:08 crc kubenswrapper[4944]: E1124 10:16:08.729455 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc26f34c-b7b7-4109-aef0-2792d712db47" containerName="mariadb-client" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.729585 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc26f34c-b7b7-4109-aef0-2792d712db47" containerName="mariadb-client" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.729888 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc26f34c-b7b7-4109-aef0-2792d712db47" containerName="mariadb-client" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.730917 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.735409 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-bhrr6" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.735787 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.736273 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.749688 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.755894 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.757540 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.761440 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.763105 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.772614 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.778555 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.892263 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xddp4\" (UniqueName: \"kubernetes.io/projected/65f35548-9555-4344-8f6a-0fe24432e01d-kube-api-access-xddp4\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.892639 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198d2038-005c-4e4d-9f85-dfa252b98986-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.892778 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/65f35548-9555-4344-8f6a-0fe24432e01d-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.892911 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-68367b44-5ae5-4293-bd09-fb831acfbe75\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68367b44-5ae5-4293-bd09-fb831acfbe75\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.892996 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65f35548-9555-4344-8f6a-0fe24432e01d-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893084 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f35548-9555-4344-8f6a-0fe24432e01d-config\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893119 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2m2k\" (UniqueName: \"kubernetes.io/projected/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-kube-api-access-x2m2k\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893217 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-config\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893296 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2d5905a5-f23f-42ab-be41-6a629537c170\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2d5905a5-f23f-42ab-be41-6a629537c170\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893353 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4dp6\" (UniqueName: \"kubernetes.io/projected/198d2038-005c-4e4d-9f85-dfa252b98986-kube-api-access-k4dp6\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893416 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f35548-9555-4344-8f6a-0fe24432e01d-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893472 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893564 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/198d2038-005c-4e4d-9f85-dfa252b98986-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893601 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1424ee34-9adc-4c96-91f8-2ae0e06097df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1424ee34-9adc-4c96-91f8-2ae0e06097df\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893660 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/198d2038-005c-4e4d-9f85-dfa252b98986-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893685 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893737 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/198d2038-005c-4e4d-9f85-dfa252b98986-config\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.893763 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.942574 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.944295 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.947681 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-8bw9h" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.947954 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.955037 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.969603 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.971066 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.981524 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.989809 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.991109 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994744 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65f35548-9555-4344-8f6a-0fe24432e01d-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994783 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f35548-9555-4344-8f6a-0fe24432e01d-config\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994802 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2m2k\" (UniqueName: \"kubernetes.io/projected/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-kube-api-access-x2m2k\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994821 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-config\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994838 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2d5905a5-f23f-42ab-be41-6a629537c170\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2d5905a5-f23f-42ab-be41-6a629537c170\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994856 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4dp6\" (UniqueName: \"kubernetes.io/projected/198d2038-005c-4e4d-9f85-dfa252b98986-kube-api-access-k4dp6\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994873 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f35548-9555-4344-8f6a-0fe24432e01d-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994890 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994915 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/198d2038-005c-4e4d-9f85-dfa252b98986-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994932 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1424ee34-9adc-4c96-91f8-2ae0e06097df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1424ee34-9adc-4c96-91f8-2ae0e06097df\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994945 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/198d2038-005c-4e4d-9f85-dfa252b98986-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994960 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.994974 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/198d2038-005c-4e4d-9f85-dfa252b98986-config\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.995001 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.995031 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xddp4\" (UniqueName: \"kubernetes.io/projected/65f35548-9555-4344-8f6a-0fe24432e01d-kube-api-access-xddp4\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.995091 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198d2038-005c-4e4d-9f85-dfa252b98986-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.995115 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/65f35548-9555-4344-8f6a-0fe24432e01d-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.995138 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-68367b44-5ae5-4293-bd09-fb831acfbe75\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68367b44-5ae5-4293-bd09-fb831acfbe75\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.995553 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.996150 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/198d2038-005c-4e4d-9f85-dfa252b98986-config\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.997824 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.997973 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f35548-9555-4344-8f6a-0fe24432e01d-config\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.998103 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/198d2038-005c-4e4d-9f85-dfa252b98986-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:08 crc kubenswrapper[4944]: I1124 10:16:08.998413 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/65f35548-9555-4344-8f6a-0fe24432e01d-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.002806 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.003726 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65f35548-9555-4344-8f6a-0fe24432e01d-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.004716 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.004718 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-config\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.004735 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2d5905a5-f23f-42ab-be41-6a629537c170\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2d5905a5-f23f-42ab-be41-6a629537c170\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f2c2a1ce72a5f51b949cab8926fa924f33976b93f3e3fe4d8075a23e046484f9/globalmount\"" pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.005347 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/198d2038-005c-4e4d-9f85-dfa252b98986-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.007984 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.008025 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-68367b44-5ae5-4293-bd09-fb831acfbe75\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68367b44-5ae5-4293-bd09-fb831acfbe75\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8830b3527f7298abafe9438e29ab66d2c4217a3e1868a97619f12aaf91cc1da6/globalmount\"" pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.009634 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.009661 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1424ee34-9adc-4c96-91f8-2ae0e06097df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1424ee34-9adc-4c96-91f8-2ae0e06097df\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/16aa159a0beda2dab6ef77ba10642d27bf8147816b6792525a8f17aaf584f25b/globalmount\"" pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.010532 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.015247 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2m2k\" (UniqueName: \"kubernetes.io/projected/9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30-kube-api-access-x2m2k\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.016210 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.027670 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198d2038-005c-4e4d-9f85-dfa252b98986-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.027991 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4dp6\" (UniqueName: \"kubernetes.io/projected/198d2038-005c-4e4d-9f85-dfa252b98986-kube-api-access-k4dp6\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.029034 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xddp4\" (UniqueName: \"kubernetes.io/projected/65f35548-9555-4344-8f6a-0fe24432e01d-kube-api-access-xddp4\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.035169 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f35548-9555-4344-8f6a-0fe24432e01d-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.074805 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-68367b44-5ae5-4293-bd09-fb831acfbe75\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68367b44-5ae5-4293-bd09-fb831acfbe75\") pod \"ovsdbserver-nb-1\" (UID: \"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30\") " pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.078312 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1424ee34-9adc-4c96-91f8-2ae0e06097df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1424ee34-9adc-4c96-91f8-2ae0e06097df\") pod \"ovsdbserver-nb-0\" (UID: \"198d2038-005c-4e4d-9f85-dfa252b98986\") " pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.080107 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.083123 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2d5905a5-f23f-42ab-be41-6a629537c170\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2d5905a5-f23f-42ab-be41-6a629537c170\") pod \"ovsdbserver-nb-2\" (UID: \"65f35548-9555-4344-8f6a-0fe24432e01d\") " pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.091759 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096106 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6db10533-47f7-48be-b9e8-2be4aa70b7c3-config\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096142 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db10533-47f7-48be-b9e8-2be4aa70b7c3-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096176 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2de0af2e-b442-490a-b850-e517d77186e5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2de0af2e-b442-490a-b850-e517d77186e5\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096196 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e421f779-63d4-4c1b-bba6-42758e256f6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e421f779-63d4-4c1b-bba6-42758e256f6c\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096281 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92de3bdd-72e5-4b71-9d2a-6279b8542747-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096319 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92de3bdd-72e5-4b71-9d2a-6279b8542747-config\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096339 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c4e65b20-fd67-4c05-8825-61b897a84543\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c4e65b20-fd67-4c05-8825-61b897a84543\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096359 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd2pj\" (UniqueName: \"kubernetes.io/projected/1543a026-7eaa-4a99-9cee-7cf077de95d0-kube-api-access-hd2pj\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096411 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77lm5\" (UniqueName: \"kubernetes.io/projected/6db10533-47f7-48be-b9e8-2be4aa70b7c3-kube-api-access-77lm5\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096599 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1543a026-7eaa-4a99-9cee-7cf077de95d0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096626 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1543a026-7eaa-4a99-9cee-7cf077de95d0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096644 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6db10533-47f7-48be-b9e8-2be4aa70b7c3-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096666 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1543a026-7eaa-4a99-9cee-7cf077de95d0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096711 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2xt6\" (UniqueName: \"kubernetes.io/projected/92de3bdd-72e5-4b71-9d2a-6279b8542747-kube-api-access-n2xt6\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096730 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92de3bdd-72e5-4b71-9d2a-6279b8542747-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096768 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/92de3bdd-72e5-4b71-9d2a-6279b8542747-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096783 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6db10533-47f7-48be-b9e8-2be4aa70b7c3-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.096799 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1543a026-7eaa-4a99-9cee-7cf077de95d0-config\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.198429 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1543a026-7eaa-4a99-9cee-7cf077de95d0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.198809 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1543a026-7eaa-4a99-9cee-7cf077de95d0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.198832 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6db10533-47f7-48be-b9e8-2be4aa70b7c3-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.198850 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1543a026-7eaa-4a99-9cee-7cf077de95d0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.198889 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2xt6\" (UniqueName: \"kubernetes.io/projected/92de3bdd-72e5-4b71-9d2a-6279b8542747-kube-api-access-n2xt6\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.198914 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92de3bdd-72e5-4b71-9d2a-6279b8542747-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.198947 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/92de3bdd-72e5-4b71-9d2a-6279b8542747-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.198965 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6db10533-47f7-48be-b9e8-2be4aa70b7c3-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.198981 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1543a026-7eaa-4a99-9cee-7cf077de95d0-config\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.199007 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6db10533-47f7-48be-b9e8-2be4aa70b7c3-config\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.199025 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db10533-47f7-48be-b9e8-2be4aa70b7c3-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.199087 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2de0af2e-b442-490a-b850-e517d77186e5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2de0af2e-b442-490a-b850-e517d77186e5\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.199113 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e421f779-63d4-4c1b-bba6-42758e256f6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e421f779-63d4-4c1b-bba6-42758e256f6c\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.199135 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92de3bdd-72e5-4b71-9d2a-6279b8542747-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.199153 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92de3bdd-72e5-4b71-9d2a-6279b8542747-config\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.199172 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c4e65b20-fd67-4c05-8825-61b897a84543\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c4e65b20-fd67-4c05-8825-61b897a84543\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.199193 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd2pj\" (UniqueName: \"kubernetes.io/projected/1543a026-7eaa-4a99-9cee-7cf077de95d0-kube-api-access-hd2pj\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.199216 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77lm5\" (UniqueName: \"kubernetes.io/projected/6db10533-47f7-48be-b9e8-2be4aa70b7c3-kube-api-access-77lm5\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.201031 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1543a026-7eaa-4a99-9cee-7cf077de95d0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.203383 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6db10533-47f7-48be-b9e8-2be4aa70b7c3-config\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.204247 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6db10533-47f7-48be-b9e8-2be4aa70b7c3-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.204581 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1543a026-7eaa-4a99-9cee-7cf077de95d0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.205887 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92de3bdd-72e5-4b71-9d2a-6279b8542747-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.206280 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/92de3bdd-72e5-4b71-9d2a-6279b8542747-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.206935 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6db10533-47f7-48be-b9e8-2be4aa70b7c3-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.207611 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1543a026-7eaa-4a99-9cee-7cf077de95d0-config\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.208514 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92de3bdd-72e5-4b71-9d2a-6279b8542747-config\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.209931 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92de3bdd-72e5-4b71-9d2a-6279b8542747-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.210860 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.210887 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c4e65b20-fd67-4c05-8825-61b897a84543\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c4e65b20-fd67-4c05-8825-61b897a84543\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b44784eca626fc744a6983cf8a712b32d18a235b4e85093937f8628123087c7e/globalmount\"" pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.216044 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.216094 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e421f779-63d4-4c1b-bba6-42758e256f6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e421f779-63d4-4c1b-bba6-42758e256f6c\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/56da08ab2d58c090e50561d1cd245855419768ddaf5a961dd73e45425d2cbecd/globalmount\"" pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.216141 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1543a026-7eaa-4a99-9cee-7cf077de95d0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.217979 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.218017 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2de0af2e-b442-490a-b850-e517d77186e5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2de0af2e-b442-490a-b850-e517d77186e5\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/145ace1958d67c3845e65d9a235939f1321334f7c7f7961057905cdd07746c8e/globalmount\"" pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.219026 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77lm5\" (UniqueName: \"kubernetes.io/projected/6db10533-47f7-48be-b9e8-2be4aa70b7c3-kube-api-access-77lm5\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.220034 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6db10533-47f7-48be-b9e8-2be4aa70b7c3-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.222765 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2xt6\" (UniqueName: \"kubernetes.io/projected/92de3bdd-72e5-4b71-9d2a-6279b8542747-kube-api-access-n2xt6\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.228633 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd2pj\" (UniqueName: \"kubernetes.io/projected/1543a026-7eaa-4a99-9cee-7cf077de95d0-kube-api-access-hd2pj\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.260556 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e421f779-63d4-4c1b-bba6-42758e256f6c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e421f779-63d4-4c1b-bba6-42758e256f6c\") pod \"ovsdbserver-sb-1\" (UID: \"6db10533-47f7-48be-b9e8-2be4aa70b7c3\") " pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.261720 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c4e65b20-fd67-4c05-8825-61b897a84543\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c4e65b20-fd67-4c05-8825-61b897a84543\") pod \"ovsdbserver-sb-0\" (UID: \"1543a026-7eaa-4a99-9cee-7cf077de95d0\") " pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.263134 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2de0af2e-b442-490a-b850-e517d77186e5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2de0af2e-b442-490a-b850-e517d77186e5\") pod \"ovsdbserver-sb-2\" (UID: \"92de3bdd-72e5-4b71-9d2a-6279b8542747\") " pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.269238 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.296339 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.314488 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.357791 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.414485 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.662243 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.816475 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 10:16:09 crc kubenswrapper[4944]: W1124 10:16:09.822401 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1543a026_7eaa_4a99_9cee_7cf077de95d0.slice/crio-8d1f1d9f8d067d06643db3041f4d6c4b3bcdd477f2eefcc80310ba56a0932890 WatchSource:0}: Error finding container 8d1f1d9f8d067d06643db3041f4d6c4b3bcdd477f2eefcc80310ba56a0932890: Status 404 returned error can't find the container with id 8d1f1d9f8d067d06643db3041f4d6c4b3bcdd477f2eefcc80310ba56a0932890 Nov 24 10:16:09 crc kubenswrapper[4944]: I1124 10:16:09.979082 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 24 10:16:10 crc kubenswrapper[4944]: I1124 10:16:10.038557 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30","Type":"ContainerStarted","Data":"b9bc418d9174130674295e4c403ea6f864bdb35a29ca6497ee4f5e6560c36bca"} Nov 24 10:16:10 crc kubenswrapper[4944]: I1124 10:16:10.038770 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30","Type":"ContainerStarted","Data":"6a557a8a8b8339c00f329c574f8bcaaebfcebdaca6556d91ba2b44926048125a"} Nov 24 10:16:10 crc kubenswrapper[4944]: I1124 10:16:10.038853 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30","Type":"ContainerStarted","Data":"cf7724d9ffae5e5c8e6031804757fd50b6a709c585a007198514111b469ea1c2"} Nov 24 10:16:10 crc kubenswrapper[4944]: I1124 10:16:10.058413 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1543a026-7eaa-4a99-9cee-7cf077de95d0","Type":"ContainerStarted","Data":"8d1f1d9f8d067d06643db3041f4d6c4b3bcdd477f2eefcc80310ba56a0932890"} Nov 24 10:16:10 crc kubenswrapper[4944]: I1124 10:16:10.068409 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"65f35548-9555-4344-8f6a-0fe24432e01d","Type":"ContainerStarted","Data":"6b6bc1e4a04938da790b8b68cd710dee6a2c70da3fa5ff026622dab0f5fae56f"} Nov 24 10:16:10 crc kubenswrapper[4944]: I1124 10:16:10.068456 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"65f35548-9555-4344-8f6a-0fe24432e01d","Type":"ContainerStarted","Data":"276bf9fd0ba28339c9ef9569dd650a45416a300ba9d366d13fa3ef976903da6c"} Nov 24 10:16:10 crc kubenswrapper[4944]: I1124 10:16:10.101099 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.101077211 podStartE2EDuration="3.101077211s" podCreationTimestamp="2025-11-24 10:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:10.072533112 +0000 UTC m=+5030.606973564" watchObservedRunningTime="2025-11-24 10:16:10.101077211 +0000 UTC m=+5030.635517673" Nov 24 10:16:10 crc kubenswrapper[4944]: I1124 10:16:10.102509 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.102504216 podStartE2EDuration="3.102504216s" podCreationTimestamp="2025-11-24 10:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:10.097671193 +0000 UTC m=+5030.632111655" watchObservedRunningTime="2025-11-24 10:16:10.102504216 +0000 UTC m=+5030.636944668" Nov 24 10:16:10 crc kubenswrapper[4944]: I1124 10:16:10.141192 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 10:16:10 crc kubenswrapper[4944]: W1124 10:16:10.148427 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod198d2038_005c_4e4d_9f85_dfa252b98986.slice/crio-fc1370b085e803b3879fe4b85f4262cde468cd7fe8319e6de42415446954f091 WatchSource:0}: Error finding container fc1370b085e803b3879fe4b85f4262cde468cd7fe8319e6de42415446954f091: Status 404 returned error can't find the container with id fc1370b085e803b3879fe4b85f4262cde468cd7fe8319e6de42415446954f091 Nov 24 10:16:10 crc kubenswrapper[4944]: I1124 10:16:10.883941 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 24 10:16:10 crc kubenswrapper[4944]: W1124 10:16:10.886868 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6db10533_47f7_48be_b9e8_2be4aa70b7c3.slice/crio-068d7160891a83be5a766ce466ecb05cea48c54af347536ae22a7736e6a97236 WatchSource:0}: Error finding container 068d7160891a83be5a766ce466ecb05cea48c54af347536ae22a7736e6a97236: Status 404 returned error can't find the container with id 068d7160891a83be5a766ce466ecb05cea48c54af347536ae22a7736e6a97236 Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.098113 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1543a026-7eaa-4a99-9cee-7cf077de95d0","Type":"ContainerStarted","Data":"9e1c0d003916d3250251ddf04cbc5d9a1a80421cf037bc88af5ff571f88b17c6"} Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.098565 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1543a026-7eaa-4a99-9cee-7cf077de95d0","Type":"ContainerStarted","Data":"d9c6433e66c14b329921d0bf7cc7b3aab7f9788c38b919433a6721ea913042b7"} Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.100076 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"65f35548-9555-4344-8f6a-0fe24432e01d","Type":"ContainerStarted","Data":"8482906354bc56e7fc354a3f7eb21d11460dec251f286c64ca8b5c8d2834cd5b"} Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.101846 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"92de3bdd-72e5-4b71-9d2a-6279b8542747","Type":"ContainerStarted","Data":"a2b8d7ebb0866eff5dcbe7e2b8488afbc366f388be4cef732fee7764935055ec"} Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.101896 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"92de3bdd-72e5-4b71-9d2a-6279b8542747","Type":"ContainerStarted","Data":"fb4998bf28e73c6fe81ff5ed040d8280d7b73fa020e659c15a862140ab22a73b"} Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.101910 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"92de3bdd-72e5-4b71-9d2a-6279b8542747","Type":"ContainerStarted","Data":"947a917bc5d5a7069e0d89db32c1851d1a72a9bad5eccba1d8ff6a58fbf61b37"} Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.103418 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"198d2038-005c-4e4d-9f85-dfa252b98986","Type":"ContainerStarted","Data":"f5306bd3dddc9e309f6a8a6d1337672eeeb9a66234e1eae63a8949cee24cdace"} Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.103444 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"198d2038-005c-4e4d-9f85-dfa252b98986","Type":"ContainerStarted","Data":"ec42dc390fb368ab69c09e01faefa3aaa96e59f69a5e1193a110e8a120a74cb2"} Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.103454 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"198d2038-005c-4e4d-9f85-dfa252b98986","Type":"ContainerStarted","Data":"fc1370b085e803b3879fe4b85f4262cde468cd7fe8319e6de42415446954f091"} Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.105193 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"6db10533-47f7-48be-b9e8-2be4aa70b7c3","Type":"ContainerStarted","Data":"dc03e7b078752b19154bd07387d62993115a0469fd6d7de36b185fa03aba6cf2"} Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.105219 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"6db10533-47f7-48be-b9e8-2be4aa70b7c3","Type":"ContainerStarted","Data":"068d7160891a83be5a766ce466ecb05cea48c54af347536ae22a7736e6a97236"} Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.136130 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.136111218 podStartE2EDuration="4.136111218s" podCreationTimestamp="2025-11-24 10:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:11.117222077 +0000 UTC m=+5031.651662539" watchObservedRunningTime="2025-11-24 10:16:11.136111218 +0000 UTC m=+5031.670551680" Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.137841 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.137835773 podStartE2EDuration="4.137835773s" podCreationTimestamp="2025-11-24 10:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:11.133380222 +0000 UTC m=+5031.667820684" watchObservedRunningTime="2025-11-24 10:16:11.137835773 +0000 UTC m=+5031.672276235" Nov 24 10:16:11 crc kubenswrapper[4944]: I1124 10:16:11.151682 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=4.151662843 podStartE2EDuration="4.151662843s" podCreationTimestamp="2025-11-24 10:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:11.14841225 +0000 UTC m=+5031.682852712" watchObservedRunningTime="2025-11-24 10:16:11.151662843 +0000 UTC m=+5031.686103305" Nov 24 10:16:12 crc kubenswrapper[4944]: I1124 10:16:12.081165 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:12 crc kubenswrapper[4944]: I1124 10:16:12.092003 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:12 crc kubenswrapper[4944]: I1124 10:16:12.114207 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"6db10533-47f7-48be-b9e8-2be4aa70b7c3","Type":"ContainerStarted","Data":"d5bdf5d1adc79f97d9c000deb304cc0217ff7670aa9a45daa0cacefc6e7b6f5d"} Nov 24 10:16:12 crc kubenswrapper[4944]: I1124 10:16:12.129495 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:12 crc kubenswrapper[4944]: I1124 10:16:12.133355 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=5.133328282 podStartE2EDuration="5.133328282s" podCreationTimestamp="2025-11-24 10:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:12.130286915 +0000 UTC m=+5032.664727417" watchObservedRunningTime="2025-11-24 10:16:12.133328282 +0000 UTC m=+5032.667768784" Nov 24 10:16:12 crc kubenswrapper[4944]: I1124 10:16:12.138565 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:12 crc kubenswrapper[4944]: I1124 10:16:12.269644 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:12 crc kubenswrapper[4944]: I1124 10:16:12.297155 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:12 crc kubenswrapper[4944]: I1124 10:16:12.315480 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:12 crc kubenswrapper[4944]: I1124 10:16:12.357863 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:13 crc kubenswrapper[4944]: I1124 10:16:13.121630 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:13 crc kubenswrapper[4944]: I1124 10:16:13.122091 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.125129 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.135764 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.270238 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.297014 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.314592 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.347903 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55bb96f665-4shk7"] Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.351388 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55bb96f665-4shk7"] Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.351559 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.359563 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.362240 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.418199 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-config\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.418248 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htv8l\" (UniqueName: \"kubernetes.io/projected/bf963104-57ae-477a-bbc9-5c93ed77ddb4-kube-api-access-htv8l\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.418292 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-dns-svc\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.418336 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-ovsdbserver-nb\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.519838 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-config\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.520352 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htv8l\" (UniqueName: \"kubernetes.io/projected/bf963104-57ae-477a-bbc9-5c93ed77ddb4-kube-api-access-htv8l\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.520691 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-dns-svc\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.521490 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-ovsdbserver-nb\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.521426 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-dns-svc\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.521035 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-config\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.522578 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-ovsdbserver-nb\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.546596 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htv8l\" (UniqueName: \"kubernetes.io/projected/bf963104-57ae-477a-bbc9-5c93ed77ddb4-kube-api-access-htv8l\") pod \"dnsmasq-dns-55bb96f665-4shk7\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:14 crc kubenswrapper[4944]: I1124 10:16:14.680943 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.093852 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55bb96f665-4shk7"] Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.143659 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" event={"ID":"bf963104-57ae-477a-bbc9-5c93ed77ddb4","Type":"ContainerStarted","Data":"6a3e46ab41aea288eb5159712b2f2c6ccac424725cf980383352b4193920edd4"} Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.317243 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.344890 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.360149 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.368015 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.392893 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.401384 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.451469 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.595010 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55bb96f665-4shk7"] Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.620539 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-678c494fd7-ggrmn"] Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.630636 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.635955 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-678c494fd7-ggrmn"] Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.639997 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.738037 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-nb\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.738120 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-dns-svc\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.738152 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnjm7\" (UniqueName: \"kubernetes.io/projected/ff5bbccb-f586-4bf5-9fc8-5579eb548278-kube-api-access-wnjm7\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.738180 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-sb\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.738368 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-config\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.839347 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-dns-svc\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.839405 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnjm7\" (UniqueName: \"kubernetes.io/projected/ff5bbccb-f586-4bf5-9fc8-5579eb548278-kube-api-access-wnjm7\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.839435 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-sb\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.839469 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-config\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.839529 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-nb\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.840370 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-dns-svc\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.840448 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-nb\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.840509 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-sb\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.840525 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-config\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.856714 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnjm7\" (UniqueName: \"kubernetes.io/projected/ff5bbccb-f586-4bf5-9fc8-5579eb548278-kube-api-access-wnjm7\") pod \"dnsmasq-dns-678c494fd7-ggrmn\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:15 crc kubenswrapper[4944]: I1124 10:16:15.947439 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:16 crc kubenswrapper[4944]: I1124 10:16:16.151368 4944 generic.go:334] "Generic (PLEG): container finished" podID="bf963104-57ae-477a-bbc9-5c93ed77ddb4" containerID="8435d6b1d21477f9a26bd1f0808bd1ca9b3c0620b66c9c40fc914b5d4ef2e38c" exitCode=0 Nov 24 10:16:16 crc kubenswrapper[4944]: I1124 10:16:16.151418 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" event={"ID":"bf963104-57ae-477a-bbc9-5c93ed77ddb4","Type":"ContainerDied","Data":"8435d6b1d21477f9a26bd1f0808bd1ca9b3c0620b66c9c40fc914b5d4ef2e38c"} Nov 24 10:16:16 crc kubenswrapper[4944]: I1124 10:16:16.199802 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Nov 24 10:16:16 crc kubenswrapper[4944]: I1124 10:16:16.277503 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:16:16 crc kubenswrapper[4944]: E1124 10:16:16.277771 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:16:16 crc kubenswrapper[4944]: I1124 10:16:16.356040 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-678c494fd7-ggrmn"] Nov 24 10:16:16 crc kubenswrapper[4944]: W1124 10:16:16.364627 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff5bbccb_f586_4bf5_9fc8_5579eb548278.slice/crio-6fca2a2237c97de735bf529791226e74fc148d2f3253659f3e9391c89aaa754a WatchSource:0}: Error finding container 6fca2a2237c97de735bf529791226e74fc148d2f3253659f3e9391c89aaa754a: Status 404 returned error can't find the container with id 6fca2a2237c97de735bf529791226e74fc148d2f3253659f3e9391c89aaa754a Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.163897 4944 generic.go:334] "Generic (PLEG): container finished" podID="ff5bbccb-f586-4bf5-9fc8-5579eb548278" containerID="3b7f0f932758a488d457bbd0ecc35528d9eb5c141d69ecce07ad18d0eaf5bfed" exitCode=0 Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.164008 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" event={"ID":"ff5bbccb-f586-4bf5-9fc8-5579eb548278","Type":"ContainerDied","Data":"3b7f0f932758a488d457bbd0ecc35528d9eb5c141d69ecce07ad18d0eaf5bfed"} Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.164408 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" event={"ID":"ff5bbccb-f586-4bf5-9fc8-5579eb548278","Type":"ContainerStarted","Data":"6fca2a2237c97de735bf529791226e74fc148d2f3253659f3e9391c89aaa754a"} Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.171467 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" event={"ID":"bf963104-57ae-477a-bbc9-5c93ed77ddb4","Type":"ContainerStarted","Data":"91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae"} Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.171528 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" podUID="bf963104-57ae-477a-bbc9-5c93ed77ddb4" containerName="dnsmasq-dns" containerID="cri-o://91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae" gracePeriod=10 Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.171607 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.226787 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" podStartSLOduration=3.226766297 podStartE2EDuration="3.226766297s" podCreationTimestamp="2025-11-24 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:17.222691777 +0000 UTC m=+5037.757132239" watchObservedRunningTime="2025-11-24 10:16:17.226766297 +0000 UTC m=+5037.761206769" Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.511534 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.667740 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htv8l\" (UniqueName: \"kubernetes.io/projected/bf963104-57ae-477a-bbc9-5c93ed77ddb4-kube-api-access-htv8l\") pod \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.667786 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-config\") pod \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.667829 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-ovsdbserver-nb\") pod \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.667875 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-dns-svc\") pod \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\" (UID: \"bf963104-57ae-477a-bbc9-5c93ed77ddb4\") " Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.681472 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf963104-57ae-477a-bbc9-5c93ed77ddb4-kube-api-access-htv8l" (OuterVolumeSpecName: "kube-api-access-htv8l") pod "bf963104-57ae-477a-bbc9-5c93ed77ddb4" (UID: "bf963104-57ae-477a-bbc9-5c93ed77ddb4"). InnerVolumeSpecName "kube-api-access-htv8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.706056 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-config" (OuterVolumeSpecName: "config") pod "bf963104-57ae-477a-bbc9-5c93ed77ddb4" (UID: "bf963104-57ae-477a-bbc9-5c93ed77ddb4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.713867 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bf963104-57ae-477a-bbc9-5c93ed77ddb4" (UID: "bf963104-57ae-477a-bbc9-5c93ed77ddb4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.721450 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bf963104-57ae-477a-bbc9-5c93ed77ddb4" (UID: "bf963104-57ae-477a-bbc9-5c93ed77ddb4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.770024 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htv8l\" (UniqueName: \"kubernetes.io/projected/bf963104-57ae-477a-bbc9-5c93ed77ddb4-kube-api-access-htv8l\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.770365 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.770376 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:17 crc kubenswrapper[4944]: I1124 10:16:17.770387 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf963104-57ae-477a-bbc9-5c93ed77ddb4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.178227 4944 generic.go:334] "Generic (PLEG): container finished" podID="bf963104-57ae-477a-bbc9-5c93ed77ddb4" containerID="91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae" exitCode=0 Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.178266 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.178302 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" event={"ID":"bf963104-57ae-477a-bbc9-5c93ed77ddb4","Type":"ContainerDied","Data":"91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae"} Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.178338 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55bb96f665-4shk7" event={"ID":"bf963104-57ae-477a-bbc9-5c93ed77ddb4","Type":"ContainerDied","Data":"6a3e46ab41aea288eb5159712b2f2c6ccac424725cf980383352b4193920edd4"} Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.178353 4944 scope.go:117] "RemoveContainer" containerID="91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.180289 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" event={"ID":"ff5bbccb-f586-4bf5-9fc8-5579eb548278","Type":"ContainerStarted","Data":"a4b386881cb0e5796bea4dfee96632381ecf94287bbc539f1958a645e9737fb5"} Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.180469 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.196696 4944 scope.go:117] "RemoveContainer" containerID="8435d6b1d21477f9a26bd1f0808bd1ca9b3c0620b66c9c40fc914b5d4ef2e38c" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.218935 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" podStartSLOduration=3.218917669 podStartE2EDuration="3.218917669s" podCreationTimestamp="2025-11-24 10:16:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:18.203170537 +0000 UTC m=+5038.737611019" watchObservedRunningTime="2025-11-24 10:16:18.218917669 +0000 UTC m=+5038.753358131" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.221568 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55bb96f665-4shk7"] Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.222474 4944 scope.go:117] "RemoveContainer" containerID="91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae" Nov 24 10:16:18 crc kubenswrapper[4944]: E1124 10:16:18.223443 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae\": container with ID starting with 91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae not found: ID does not exist" containerID="91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.223479 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae"} err="failed to get container status \"91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae\": rpc error: code = NotFound desc = could not find container \"91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae\": container with ID starting with 91f99d6bb33593e793fabf42a6e1785e6a78083c64ff547fccbfa22dc7a2e1ae not found: ID does not exist" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.223503 4944 scope.go:117] "RemoveContainer" containerID="8435d6b1d21477f9a26bd1f0808bd1ca9b3c0620b66c9c40fc914b5d4ef2e38c" Nov 24 10:16:18 crc kubenswrapper[4944]: E1124 10:16:18.223768 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8435d6b1d21477f9a26bd1f0808bd1ca9b3c0620b66c9c40fc914b5d4ef2e38c\": container with ID starting with 8435d6b1d21477f9a26bd1f0808bd1ca9b3c0620b66c9c40fc914b5d4ef2e38c not found: ID does not exist" containerID="8435d6b1d21477f9a26bd1f0808bd1ca9b3c0620b66c9c40fc914b5d4ef2e38c" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.223794 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8435d6b1d21477f9a26bd1f0808bd1ca9b3c0620b66c9c40fc914b5d4ef2e38c"} err="failed to get container status \"8435d6b1d21477f9a26bd1f0808bd1ca9b3c0620b66c9c40fc914b5d4ef2e38c\": rpc error: code = NotFound desc = could not find container \"8435d6b1d21477f9a26bd1f0808bd1ca9b3c0620b66c9c40fc914b5d4ef2e38c\": container with ID starting with 8435d6b1d21477f9a26bd1f0808bd1ca9b3c0620b66c9c40fc914b5d4ef2e38c not found: ID does not exist" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.226449 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55bb96f665-4shk7"] Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.289336 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf963104-57ae-477a-bbc9-5c93ed77ddb4" path="/var/lib/kubelet/pods/bf963104-57ae-477a-bbc9-5c93ed77ddb4/volumes" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.594032 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Nov 24 10:16:18 crc kubenswrapper[4944]: E1124 10:16:18.594498 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf963104-57ae-477a-bbc9-5c93ed77ddb4" containerName="dnsmasq-dns" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.594526 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf963104-57ae-477a-bbc9-5c93ed77ddb4" containerName="dnsmasq-dns" Nov 24 10:16:18 crc kubenswrapper[4944]: E1124 10:16:18.594573 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf963104-57ae-477a-bbc9-5c93ed77ddb4" containerName="init" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.594582 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf963104-57ae-477a-bbc9-5c93ed77ddb4" containerName="init" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.594780 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf963104-57ae-477a-bbc9-5c93ed77ddb4" containerName="dnsmasq-dns" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.595623 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.597350 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.600840 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.682791 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gcnz\" (UniqueName: \"kubernetes.io/projected/9c64cab4-f4a0-4800-b401-3f3d107dfce1-kube-api-access-4gcnz\") pod \"ovn-copy-data\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " pod="openstack/ovn-copy-data" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.682857 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-04716da7-9e6d-4a95-810d-86d54b500df3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04716da7-9e6d-4a95-810d-86d54b500df3\") pod \"ovn-copy-data\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " pod="openstack/ovn-copy-data" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.682940 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/9c64cab4-f4a0-4800-b401-3f3d107dfce1-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " pod="openstack/ovn-copy-data" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.784286 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/9c64cab4-f4a0-4800-b401-3f3d107dfce1-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " pod="openstack/ovn-copy-data" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.784368 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gcnz\" (UniqueName: \"kubernetes.io/projected/9c64cab4-f4a0-4800-b401-3f3d107dfce1-kube-api-access-4gcnz\") pod \"ovn-copy-data\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " pod="openstack/ovn-copy-data" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.784402 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-04716da7-9e6d-4a95-810d-86d54b500df3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04716da7-9e6d-4a95-810d-86d54b500df3\") pod \"ovn-copy-data\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " pod="openstack/ovn-copy-data" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.788596 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.788648 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-04716da7-9e6d-4a95-810d-86d54b500df3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04716da7-9e6d-4a95-810d-86d54b500df3\") pod \"ovn-copy-data\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f0f104a10dbec5cf9ce06fa38f9b1367cc5286c49f68c1a023e251fd2012ceba/globalmount\"" pod="openstack/ovn-copy-data" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.789984 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/9c64cab4-f4a0-4800-b401-3f3d107dfce1-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " pod="openstack/ovn-copy-data" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.801392 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gcnz\" (UniqueName: \"kubernetes.io/projected/9c64cab4-f4a0-4800-b401-3f3d107dfce1-kube-api-access-4gcnz\") pod \"ovn-copy-data\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " pod="openstack/ovn-copy-data" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.816253 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-04716da7-9e6d-4a95-810d-86d54b500df3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04716da7-9e6d-4a95-810d-86d54b500df3\") pod \"ovn-copy-data\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " pod="openstack/ovn-copy-data" Nov 24 10:16:18 crc kubenswrapper[4944]: I1124 10:16:18.973588 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 10:16:19 crc kubenswrapper[4944]: I1124 10:16:19.234981 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 10:16:19 crc kubenswrapper[4944]: W1124 10:16:19.245193 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c64cab4_f4a0_4800_b401_3f3d107dfce1.slice/crio-7ee0f4d42d27fdd1c3967c411e735675a0565701b6581515c2b6a66a181e203c WatchSource:0}: Error finding container 7ee0f4d42d27fdd1c3967c411e735675a0565701b6581515c2b6a66a181e203c: Status 404 returned error can't find the container with id 7ee0f4d42d27fdd1c3967c411e735675a0565701b6581515c2b6a66a181e203c Nov 24 10:16:20 crc kubenswrapper[4944]: I1124 10:16:20.201211 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"9c64cab4-f4a0-4800-b401-3f3d107dfce1","Type":"ContainerStarted","Data":"97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745"} Nov 24 10:16:20 crc kubenswrapper[4944]: I1124 10:16:20.201562 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"9c64cab4-f4a0-4800-b401-3f3d107dfce1","Type":"ContainerStarted","Data":"7ee0f4d42d27fdd1c3967c411e735675a0565701b6581515c2b6a66a181e203c"} Nov 24 10:16:20 crc kubenswrapper[4944]: I1124 10:16:20.222558 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.222542078 podStartE2EDuration="3.222542078s" podCreationTimestamp="2025-11-24 10:16:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:20.217648592 +0000 UTC m=+5040.752089074" watchObservedRunningTime="2025-11-24 10:16:20.222542078 +0000 UTC m=+5040.756982540" Nov 24 10:16:24 crc kubenswrapper[4944]: I1124 10:16:24.808081 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 10:16:24 crc kubenswrapper[4944]: I1124 10:16:24.810917 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 10:16:24 crc kubenswrapper[4944]: I1124 10:16:24.818442 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 10:16:24 crc kubenswrapper[4944]: I1124 10:16:24.818759 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 10:16:24 crc kubenswrapper[4944]: I1124 10:16:24.818855 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-cdzb6" Nov 24 10:16:24 crc kubenswrapper[4944]: I1124 10:16:24.832431 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 10:16:24 crc kubenswrapper[4944]: I1124 10:16:24.987337 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-config\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:24 crc kubenswrapper[4944]: I1124 10:16:24.987384 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27tqj\" (UniqueName: \"kubernetes.io/projected/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-kube-api-access-27tqj\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:24 crc kubenswrapper[4944]: I1124 10:16:24.987415 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:24 crc kubenswrapper[4944]: I1124 10:16:24.987468 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-scripts\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:24 crc kubenswrapper[4944]: I1124 10:16:24.987532 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.088728 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-config\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.088786 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27tqj\" (UniqueName: \"kubernetes.io/projected/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-kube-api-access-27tqj\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.088829 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.088920 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-scripts\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.089038 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.089688 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.089994 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-scripts\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.090646 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-config\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.095111 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.108973 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27tqj\" (UniqueName: \"kubernetes.io/projected/561e45c8-64b9-4eef-a8a9-b1620e1d03b1-kube-api-access-27tqj\") pod \"ovn-northd-0\" (UID: \"561e45c8-64b9-4eef-a8a9-b1620e1d03b1\") " pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.134105 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.621397 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 10:16:25 crc kubenswrapper[4944]: W1124 10:16:25.632827 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod561e45c8_64b9_4eef_a8a9_b1620e1d03b1.slice/crio-08a5d5744609225e0af288d2fda04ae6046a1dce633537813198cdf7659d2477 WatchSource:0}: Error finding container 08a5d5744609225e0af288d2fda04ae6046a1dce633537813198cdf7659d2477: Status 404 returned error can't find the container with id 08a5d5744609225e0af288d2fda04ae6046a1dce633537813198cdf7659d2477 Nov 24 10:16:25 crc kubenswrapper[4944]: I1124 10:16:25.948960 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.004388 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-png4d"] Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.004665 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" podUID="7ee58980-1615-4769-8d7b-fda64bd2276c" containerName="dnsmasq-dns" containerID="cri-o://31c0689fd589e8d8dd1d24c1f5ebd0d951cf9e366b99499a678950a3826e2f6d" gracePeriod=10 Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.261460 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"561e45c8-64b9-4eef-a8a9-b1620e1d03b1","Type":"ContainerStarted","Data":"283c0406dae129233c7e78db493e066042c0eb2d0eb568f00e3aceb5c3eb4cd5"} Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.262099 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"561e45c8-64b9-4eef-a8a9-b1620e1d03b1","Type":"ContainerStarted","Data":"e9caa641aac9ebcb6046257a96bd1306623263b7b736f9455dfbd95d3160312e"} Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.262114 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"561e45c8-64b9-4eef-a8a9-b1620e1d03b1","Type":"ContainerStarted","Data":"08a5d5744609225e0af288d2fda04ae6046a1dce633537813198cdf7659d2477"} Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.262246 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.264124 4944 generic.go:334] "Generic (PLEG): container finished" podID="7ee58980-1615-4769-8d7b-fda64bd2276c" containerID="31c0689fd589e8d8dd1d24c1f5ebd0d951cf9e366b99499a678950a3826e2f6d" exitCode=0 Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.264164 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" event={"ID":"7ee58980-1615-4769-8d7b-fda64bd2276c","Type":"ContainerDied","Data":"31c0689fd589e8d8dd1d24c1f5ebd0d951cf9e366b99499a678950a3826e2f6d"} Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.289255 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.289230872 podStartE2EDuration="2.289230872s" podCreationTimestamp="2025-11-24 10:16:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:26.280708891 +0000 UTC m=+5046.815149373" watchObservedRunningTime="2025-11-24 10:16:26.289230872 +0000 UTC m=+5046.823671354" Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.446860 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.520518 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-dns-svc\") pod \"7ee58980-1615-4769-8d7b-fda64bd2276c\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.521113 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-config\") pod \"7ee58980-1615-4769-8d7b-fda64bd2276c\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.521296 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2bv4\" (UniqueName: \"kubernetes.io/projected/7ee58980-1615-4769-8d7b-fda64bd2276c-kube-api-access-c2bv4\") pod \"7ee58980-1615-4769-8d7b-fda64bd2276c\" (UID: \"7ee58980-1615-4769-8d7b-fda64bd2276c\") " Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.525017 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee58980-1615-4769-8d7b-fda64bd2276c-kube-api-access-c2bv4" (OuterVolumeSpecName: "kube-api-access-c2bv4") pod "7ee58980-1615-4769-8d7b-fda64bd2276c" (UID: "7ee58980-1615-4769-8d7b-fda64bd2276c"). InnerVolumeSpecName "kube-api-access-c2bv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.556008 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-config" (OuterVolumeSpecName: "config") pod "7ee58980-1615-4769-8d7b-fda64bd2276c" (UID: "7ee58980-1615-4769-8d7b-fda64bd2276c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.556571 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7ee58980-1615-4769-8d7b-fda64bd2276c" (UID: "7ee58980-1615-4769-8d7b-fda64bd2276c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.623965 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2bv4\" (UniqueName: \"kubernetes.io/projected/7ee58980-1615-4769-8d7b-fda64bd2276c-kube-api-access-c2bv4\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.623994 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:26 crc kubenswrapper[4944]: I1124 10:16:26.624003 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ee58980-1615-4769-8d7b-fda64bd2276c-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:27 crc kubenswrapper[4944]: I1124 10:16:27.274291 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" event={"ID":"7ee58980-1615-4769-8d7b-fda64bd2276c","Type":"ContainerDied","Data":"1e51c371b0ca6f8e67680bffd9c155d00cf12a1f5c9fd9738cb686e64bc1077d"} Nov 24 10:16:27 crc kubenswrapper[4944]: I1124 10:16:27.274720 4944 scope.go:117] "RemoveContainer" containerID="31c0689fd589e8d8dd1d24c1f5ebd0d951cf9e366b99499a678950a3826e2f6d" Nov 24 10:16:27 crc kubenswrapper[4944]: I1124 10:16:27.274343 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-png4d" Nov 24 10:16:27 crc kubenswrapper[4944]: I1124 10:16:27.293539 4944 scope.go:117] "RemoveContainer" containerID="636fda1fcd9206ae51ad46629bff26db37ab8d4aa56be5818e3747d3bd3c9486" Nov 24 10:16:27 crc kubenswrapper[4944]: I1124 10:16:27.314499 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-png4d"] Nov 24 10:16:27 crc kubenswrapper[4944]: I1124 10:16:27.322028 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-png4d"] Nov 24 10:16:28 crc kubenswrapper[4944]: I1124 10:16:28.288183 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ee58980-1615-4769-8d7b-fda64bd2276c" path="/var/lib/kubelet/pods/7ee58980-1615-4769-8d7b-fda64bd2276c/volumes" Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.835324 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-nbr9v"] Nov 24 10:16:29 crc kubenswrapper[4944]: E1124 10:16:29.835996 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee58980-1615-4769-8d7b-fda64bd2276c" containerName="dnsmasq-dns" Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.836008 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee58980-1615-4769-8d7b-fda64bd2276c" containerName="dnsmasq-dns" Nov 24 10:16:29 crc kubenswrapper[4944]: E1124 10:16:29.836022 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee58980-1615-4769-8d7b-fda64bd2276c" containerName="init" Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.836028 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee58980-1615-4769-8d7b-fda64bd2276c" containerName="init" Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.836186 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee58980-1615-4769-8d7b-fda64bd2276c" containerName="dnsmasq-dns" Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.836676 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nbr9v" Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.847825 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nbr9v"] Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.938668 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-812f-account-create-ksj7z"] Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.939602 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-812f-account-create-ksj7z" Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.941406 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.954143 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-812f-account-create-ksj7z"] Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.989389 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-operator-scripts\") pod \"keystone-db-create-nbr9v\" (UID: \"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b\") " pod="openstack/keystone-db-create-nbr9v" Nov 24 10:16:29 crc kubenswrapper[4944]: I1124 10:16:29.989596 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x949\" (UniqueName: \"kubernetes.io/projected/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-kube-api-access-8x949\") pod \"keystone-db-create-nbr9v\" (UID: \"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b\") " pod="openstack/keystone-db-create-nbr9v" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.091388 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-operator-scripts\") pod \"keystone-db-create-nbr9v\" (UID: \"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b\") " pod="openstack/keystone-db-create-nbr9v" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.091486 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b7743bd-f104-434a-9a4d-59db71cb9a97-operator-scripts\") pod \"keystone-812f-account-create-ksj7z\" (UID: \"4b7743bd-f104-434a-9a4d-59db71cb9a97\") " pod="openstack/keystone-812f-account-create-ksj7z" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.091551 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x949\" (UniqueName: \"kubernetes.io/projected/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-kube-api-access-8x949\") pod \"keystone-db-create-nbr9v\" (UID: \"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b\") " pod="openstack/keystone-db-create-nbr9v" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.091569 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x5jg\" (UniqueName: \"kubernetes.io/projected/4b7743bd-f104-434a-9a4d-59db71cb9a97-kube-api-access-7x5jg\") pod \"keystone-812f-account-create-ksj7z\" (UID: \"4b7743bd-f104-434a-9a4d-59db71cb9a97\") " pod="openstack/keystone-812f-account-create-ksj7z" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.092446 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-operator-scripts\") pod \"keystone-db-create-nbr9v\" (UID: \"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b\") " pod="openstack/keystone-db-create-nbr9v" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.115772 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x949\" (UniqueName: \"kubernetes.io/projected/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-kube-api-access-8x949\") pod \"keystone-db-create-nbr9v\" (UID: \"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b\") " pod="openstack/keystone-db-create-nbr9v" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.187819 4944 scope.go:117] "RemoveContainer" containerID="e9fc5f229012c10e7ad3821d40f878fef079197c3588dca3eba0338022f1480e" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.192637 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b7743bd-f104-434a-9a4d-59db71cb9a97-operator-scripts\") pod \"keystone-812f-account-create-ksj7z\" (UID: \"4b7743bd-f104-434a-9a4d-59db71cb9a97\") " pod="openstack/keystone-812f-account-create-ksj7z" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.193769 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b7743bd-f104-434a-9a4d-59db71cb9a97-operator-scripts\") pod \"keystone-812f-account-create-ksj7z\" (UID: \"4b7743bd-f104-434a-9a4d-59db71cb9a97\") " pod="openstack/keystone-812f-account-create-ksj7z" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.193934 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nbr9v" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.194340 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x5jg\" (UniqueName: \"kubernetes.io/projected/4b7743bd-f104-434a-9a4d-59db71cb9a97-kube-api-access-7x5jg\") pod \"keystone-812f-account-create-ksj7z\" (UID: \"4b7743bd-f104-434a-9a4d-59db71cb9a97\") " pod="openstack/keystone-812f-account-create-ksj7z" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.212250 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x5jg\" (UniqueName: \"kubernetes.io/projected/4b7743bd-f104-434a-9a4d-59db71cb9a97-kube-api-access-7x5jg\") pod \"keystone-812f-account-create-ksj7z\" (UID: \"4b7743bd-f104-434a-9a4d-59db71cb9a97\") " pod="openstack/keystone-812f-account-create-ksj7z" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.224844 4944 scope.go:117] "RemoveContainer" containerID="ad0b7bc15faa435934ff6851c28afc33050316559d83d902e8b8397ef76e06f7" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.256149 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-812f-account-create-ksj7z" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.368877 4944 scope.go:117] "RemoveContainer" containerID="d1d5a9d00c4e83b11041f4eb063e3b07b6cfb4aca9171c38c3c98b45a9fb10d8" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.416244 4944 scope.go:117] "RemoveContainer" containerID="aa2726ee6bfd4f4cf805453a184118501cd210493f886d7d77138f780d87753f" Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.681673 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nbr9v"] Nov 24 10:16:30 crc kubenswrapper[4944]: W1124 10:16:30.684637 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba4b8d12_578d_4a2e_836a_09c2b86b2a9b.slice/crio-4074ec5353a93424cd5950d9fa7e84143a221eafe8a09313bc145e94bba8e955 WatchSource:0}: Error finding container 4074ec5353a93424cd5950d9fa7e84143a221eafe8a09313bc145e94bba8e955: Status 404 returned error can't find the container with id 4074ec5353a93424cd5950d9fa7e84143a221eafe8a09313bc145e94bba8e955 Nov 24 10:16:30 crc kubenswrapper[4944]: W1124 10:16:30.760269 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b7743bd_f104_434a_9a4d_59db71cb9a97.slice/crio-de25e5dd45744376d3dfd05f5f20341079361623c575dd2968fe29abe3d5fe4b WatchSource:0}: Error finding container de25e5dd45744376d3dfd05f5f20341079361623c575dd2968fe29abe3d5fe4b: Status 404 returned error can't find the container with id de25e5dd45744376d3dfd05f5f20341079361623c575dd2968fe29abe3d5fe4b Nov 24 10:16:30 crc kubenswrapper[4944]: I1124 10:16:30.760820 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-812f-account-create-ksj7z"] Nov 24 10:16:31 crc kubenswrapper[4944]: I1124 10:16:31.276640 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:16:31 crc kubenswrapper[4944]: E1124 10:16:31.277025 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:16:31 crc kubenswrapper[4944]: I1124 10:16:31.335075 4944 generic.go:334] "Generic (PLEG): container finished" podID="4b7743bd-f104-434a-9a4d-59db71cb9a97" containerID="e7a1d58325c3cc36b5958c751173ec3e753c9be72c28474a0d8802843a24d0b4" exitCode=0 Nov 24 10:16:31 crc kubenswrapper[4944]: I1124 10:16:31.335147 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-812f-account-create-ksj7z" event={"ID":"4b7743bd-f104-434a-9a4d-59db71cb9a97","Type":"ContainerDied","Data":"e7a1d58325c3cc36b5958c751173ec3e753c9be72c28474a0d8802843a24d0b4"} Nov 24 10:16:31 crc kubenswrapper[4944]: I1124 10:16:31.335174 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-812f-account-create-ksj7z" event={"ID":"4b7743bd-f104-434a-9a4d-59db71cb9a97","Type":"ContainerStarted","Data":"de25e5dd45744376d3dfd05f5f20341079361623c575dd2968fe29abe3d5fe4b"} Nov 24 10:16:31 crc kubenswrapper[4944]: I1124 10:16:31.336603 4944 generic.go:334] "Generic (PLEG): container finished" podID="ba4b8d12-578d-4a2e-836a-09c2b86b2a9b" containerID="cc8f9086bedc6d81ee3f7d8d6a18104f46373ed86032765f5f842f919418018b" exitCode=0 Nov 24 10:16:31 crc kubenswrapper[4944]: I1124 10:16:31.336638 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nbr9v" event={"ID":"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b","Type":"ContainerDied","Data":"cc8f9086bedc6d81ee3f7d8d6a18104f46373ed86032765f5f842f919418018b"} Nov 24 10:16:31 crc kubenswrapper[4944]: I1124 10:16:31.336656 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nbr9v" event={"ID":"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b","Type":"ContainerStarted","Data":"4074ec5353a93424cd5950d9fa7e84143a221eafe8a09313bc145e94bba8e955"} Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.005398 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nbr9v" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.096364 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-812f-account-create-ksj7z" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.102974 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x949\" (UniqueName: \"kubernetes.io/projected/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-kube-api-access-8x949\") pod \"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b\" (UID: \"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b\") " Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.103030 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x5jg\" (UniqueName: \"kubernetes.io/projected/4b7743bd-f104-434a-9a4d-59db71cb9a97-kube-api-access-7x5jg\") pod \"4b7743bd-f104-434a-9a4d-59db71cb9a97\" (UID: \"4b7743bd-f104-434a-9a4d-59db71cb9a97\") " Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.103092 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b7743bd-f104-434a-9a4d-59db71cb9a97-operator-scripts\") pod \"4b7743bd-f104-434a-9a4d-59db71cb9a97\" (UID: \"4b7743bd-f104-434a-9a4d-59db71cb9a97\") " Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.103134 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-operator-scripts\") pod \"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b\" (UID: \"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b\") " Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.104097 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ba4b8d12-578d-4a2e-836a-09c2b86b2a9b" (UID: "ba4b8d12-578d-4a2e-836a-09c2b86b2a9b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.104652 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b7743bd-f104-434a-9a4d-59db71cb9a97-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4b7743bd-f104-434a-9a4d-59db71cb9a97" (UID: "4b7743bd-f104-434a-9a4d-59db71cb9a97"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.111892 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b7743bd-f104-434a-9a4d-59db71cb9a97-kube-api-access-7x5jg" (OuterVolumeSpecName: "kube-api-access-7x5jg") pod "4b7743bd-f104-434a-9a4d-59db71cb9a97" (UID: "4b7743bd-f104-434a-9a4d-59db71cb9a97"). InnerVolumeSpecName "kube-api-access-7x5jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.114240 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-kube-api-access-8x949" (OuterVolumeSpecName: "kube-api-access-8x949") pod "ba4b8d12-578d-4a2e-836a-09c2b86b2a9b" (UID: "ba4b8d12-578d-4a2e-836a-09c2b86b2a9b"). InnerVolumeSpecName "kube-api-access-8x949". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.204649 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x949\" (UniqueName: \"kubernetes.io/projected/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-kube-api-access-8x949\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.204688 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x5jg\" (UniqueName: \"kubernetes.io/projected/4b7743bd-f104-434a-9a4d-59db71cb9a97-kube-api-access-7x5jg\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.204699 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b7743bd-f104-434a-9a4d-59db71cb9a97-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.204707 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.353268 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-812f-account-create-ksj7z" event={"ID":"4b7743bd-f104-434a-9a4d-59db71cb9a97","Type":"ContainerDied","Data":"de25e5dd45744376d3dfd05f5f20341079361623c575dd2968fe29abe3d5fe4b"} Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.353307 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de25e5dd45744376d3dfd05f5f20341079361623c575dd2968fe29abe3d5fe4b" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.353287 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-812f-account-create-ksj7z" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.354882 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nbr9v" event={"ID":"ba4b8d12-578d-4a2e-836a-09c2b86b2a9b","Type":"ContainerDied","Data":"4074ec5353a93424cd5950d9fa7e84143a221eafe8a09313bc145e94bba8e955"} Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.354913 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4074ec5353a93424cd5950d9fa7e84143a221eafe8a09313bc145e94bba8e955" Nov 24 10:16:33 crc kubenswrapper[4944]: I1124 10:16:33.354920 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nbr9v" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.212280 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.532602 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-jm4kh"] Nov 24 10:16:35 crc kubenswrapper[4944]: E1124 10:16:35.532902 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b7743bd-f104-434a-9a4d-59db71cb9a97" containerName="mariadb-account-create" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.532914 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b7743bd-f104-434a-9a4d-59db71cb9a97" containerName="mariadb-account-create" Nov 24 10:16:35 crc kubenswrapper[4944]: E1124 10:16:35.532946 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba4b8d12-578d-4a2e-836a-09c2b86b2a9b" containerName="mariadb-database-create" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.532951 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba4b8d12-578d-4a2e-836a-09c2b86b2a9b" containerName="mariadb-database-create" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.533089 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b7743bd-f104-434a-9a4d-59db71cb9a97" containerName="mariadb-account-create" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.533108 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba4b8d12-578d-4a2e-836a-09c2b86b2a9b" containerName="mariadb-database-create" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.533608 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.535220 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxvsz" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.535514 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.535819 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.541006 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.544166 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jm4kh"] Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.639556 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-config-data\") pod \"keystone-db-sync-jm4kh\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.639728 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-combined-ca-bundle\") pod \"keystone-db-sync-jm4kh\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.639806 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf9xn\" (UniqueName: \"kubernetes.io/projected/a9675c22-64db-452a-8d38-dd588bd5b43a-kube-api-access-gf9xn\") pod \"keystone-db-sync-jm4kh\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.741717 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-config-data\") pod \"keystone-db-sync-jm4kh\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.741774 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-combined-ca-bundle\") pod \"keystone-db-sync-jm4kh\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.741826 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf9xn\" (UniqueName: \"kubernetes.io/projected/a9675c22-64db-452a-8d38-dd588bd5b43a-kube-api-access-gf9xn\") pod \"keystone-db-sync-jm4kh\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.754076 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-combined-ca-bundle\") pod \"keystone-db-sync-jm4kh\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.754613 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-config-data\") pod \"keystone-db-sync-jm4kh\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.759708 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf9xn\" (UniqueName: \"kubernetes.io/projected/a9675c22-64db-452a-8d38-dd588bd5b43a-kube-api-access-gf9xn\") pod \"keystone-db-sync-jm4kh\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:35 crc kubenswrapper[4944]: I1124 10:16:35.848645 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:36 crc kubenswrapper[4944]: I1124 10:16:36.253450 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jm4kh"] Nov 24 10:16:36 crc kubenswrapper[4944]: W1124 10:16:36.256347 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9675c22_64db_452a_8d38_dd588bd5b43a.slice/crio-dd13a4f3de6c132c6f8a29a85765404402cbcd1eab9e45a4e23985c5a329f6c7 WatchSource:0}: Error finding container dd13a4f3de6c132c6f8a29a85765404402cbcd1eab9e45a4e23985c5a329f6c7: Status 404 returned error can't find the container with id dd13a4f3de6c132c6f8a29a85765404402cbcd1eab9e45a4e23985c5a329f6c7 Nov 24 10:16:36 crc kubenswrapper[4944]: I1124 10:16:36.380484 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jm4kh" event={"ID":"a9675c22-64db-452a-8d38-dd588bd5b43a","Type":"ContainerStarted","Data":"dd13a4f3de6c132c6f8a29a85765404402cbcd1eab9e45a4e23985c5a329f6c7"} Nov 24 10:16:37 crc kubenswrapper[4944]: I1124 10:16:37.389707 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jm4kh" event={"ID":"a9675c22-64db-452a-8d38-dd588bd5b43a","Type":"ContainerStarted","Data":"e3f3e2983ce95602958fd55b450d5259f2022bf4446fd4ebb13af3b473aba9ac"} Nov 24 10:16:37 crc kubenswrapper[4944]: I1124 10:16:37.406846 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-jm4kh" podStartSLOduration=2.406830256 podStartE2EDuration="2.406830256s" podCreationTimestamp="2025-11-24 10:16:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:37.405105902 +0000 UTC m=+5057.939546364" watchObservedRunningTime="2025-11-24 10:16:37.406830256 +0000 UTC m=+5057.941270718" Nov 24 10:16:38 crc kubenswrapper[4944]: I1124 10:16:38.399458 4944 generic.go:334] "Generic (PLEG): container finished" podID="a9675c22-64db-452a-8d38-dd588bd5b43a" containerID="e3f3e2983ce95602958fd55b450d5259f2022bf4446fd4ebb13af3b473aba9ac" exitCode=0 Nov 24 10:16:38 crc kubenswrapper[4944]: I1124 10:16:38.399540 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jm4kh" event={"ID":"a9675c22-64db-452a-8d38-dd588bd5b43a","Type":"ContainerDied","Data":"e3f3e2983ce95602958fd55b450d5259f2022bf4446fd4ebb13af3b473aba9ac"} Nov 24 10:16:39 crc kubenswrapper[4944]: I1124 10:16:39.738973 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:39 crc kubenswrapper[4944]: I1124 10:16:39.912634 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-config-data\") pod \"a9675c22-64db-452a-8d38-dd588bd5b43a\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " Nov 24 10:16:39 crc kubenswrapper[4944]: I1124 10:16:39.912929 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf9xn\" (UniqueName: \"kubernetes.io/projected/a9675c22-64db-452a-8d38-dd588bd5b43a-kube-api-access-gf9xn\") pod \"a9675c22-64db-452a-8d38-dd588bd5b43a\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " Nov 24 10:16:39 crc kubenswrapper[4944]: I1124 10:16:39.913079 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-combined-ca-bundle\") pod \"a9675c22-64db-452a-8d38-dd588bd5b43a\" (UID: \"a9675c22-64db-452a-8d38-dd588bd5b43a\") " Nov 24 10:16:39 crc kubenswrapper[4944]: I1124 10:16:39.917964 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9675c22-64db-452a-8d38-dd588bd5b43a-kube-api-access-gf9xn" (OuterVolumeSpecName: "kube-api-access-gf9xn") pod "a9675c22-64db-452a-8d38-dd588bd5b43a" (UID: "a9675c22-64db-452a-8d38-dd588bd5b43a"). InnerVolumeSpecName "kube-api-access-gf9xn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:16:39 crc kubenswrapper[4944]: I1124 10:16:39.934759 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9675c22-64db-452a-8d38-dd588bd5b43a" (UID: "a9675c22-64db-452a-8d38-dd588bd5b43a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:39 crc kubenswrapper[4944]: I1124 10:16:39.954217 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-config-data" (OuterVolumeSpecName: "config-data") pod "a9675c22-64db-452a-8d38-dd588bd5b43a" (UID: "a9675c22-64db-452a-8d38-dd588bd5b43a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.015328 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.015361 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf9xn\" (UniqueName: \"kubernetes.io/projected/a9675c22-64db-452a-8d38-dd588bd5b43a-kube-api-access-gf9xn\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.015373 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9675c22-64db-452a-8d38-dd588bd5b43a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.417368 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jm4kh" event={"ID":"a9675c22-64db-452a-8d38-dd588bd5b43a","Type":"ContainerDied","Data":"dd13a4f3de6c132c6f8a29a85765404402cbcd1eab9e45a4e23985c5a329f6c7"} Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.417412 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd13a4f3de6c132c6f8a29a85765404402cbcd1eab9e45a4e23985c5a329f6c7" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.417464 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jm4kh" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.642773 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75dd7fd5b7-c8clf"] Nov 24 10:16:40 crc kubenswrapper[4944]: E1124 10:16:40.643484 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9675c22-64db-452a-8d38-dd588bd5b43a" containerName="keystone-db-sync" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.643506 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9675c22-64db-452a-8d38-dd588bd5b43a" containerName="keystone-db-sync" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.643730 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9675c22-64db-452a-8d38-dd588bd5b43a" containerName="keystone-db-sync" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.644761 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.658165 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75dd7fd5b7-c8clf"] Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.695891 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fsrcc"] Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.699204 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.704018 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.704320 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.704419 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.704508 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxvsz" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.704595 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.708960 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fsrcc"] Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.825645 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-sb\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.825822 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-config\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.825913 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-combined-ca-bundle\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.825935 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-dns-svc\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.825986 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rbr2\" (UniqueName: \"kubernetes.io/projected/d0ef4af1-b83d-4edf-99e1-58ae502aa301-kube-api-access-5rbr2\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.826018 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-config-data\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.826147 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-nb\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.826206 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-fernet-keys\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.826265 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-scripts\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.826293 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rds5\" (UniqueName: \"kubernetes.io/projected/2ee27e85-9fdc-4440-9747-ae08bf089bd3-kube-api-access-2rds5\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.826331 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-credential-keys\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.929359 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-scripts\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.929411 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rds5\" (UniqueName: \"kubernetes.io/projected/2ee27e85-9fdc-4440-9747-ae08bf089bd3-kube-api-access-2rds5\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.929441 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-credential-keys\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.929498 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-sb\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.929538 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-config\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.929568 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-combined-ca-bundle\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.929585 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-dns-svc\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.929610 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rbr2\" (UniqueName: \"kubernetes.io/projected/d0ef4af1-b83d-4edf-99e1-58ae502aa301-kube-api-access-5rbr2\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.929632 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-config-data\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.929649 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-nb\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.929677 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-fernet-keys\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.930337 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-sb\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.930559 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-dns-svc\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.930878 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-config\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.930888 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-nb\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.933930 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-fernet-keys\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.933991 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-config-data\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.934995 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-scripts\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.935121 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-combined-ca-bundle\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.937128 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-credential-keys\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.949914 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rbr2\" (UniqueName: \"kubernetes.io/projected/d0ef4af1-b83d-4edf-99e1-58ae502aa301-kube-api-access-5rbr2\") pod \"keystone-bootstrap-fsrcc\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.951393 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rds5\" (UniqueName: \"kubernetes.io/projected/2ee27e85-9fdc-4440-9747-ae08bf089bd3-kube-api-access-2rds5\") pod \"dnsmasq-dns-75dd7fd5b7-c8clf\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:40 crc kubenswrapper[4944]: I1124 10:16:40.966496 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:41 crc kubenswrapper[4944]: I1124 10:16:41.014274 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:41 crc kubenswrapper[4944]: I1124 10:16:41.409496 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75dd7fd5b7-c8clf"] Nov 24 10:16:41 crc kubenswrapper[4944]: I1124 10:16:41.433240 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" event={"ID":"2ee27e85-9fdc-4440-9747-ae08bf089bd3","Type":"ContainerStarted","Data":"beb2e81c0c788a97aee9cab316426bcae4c57190a5c80e6a13f7cbe10abc1c6d"} Nov 24 10:16:41 crc kubenswrapper[4944]: I1124 10:16:41.507293 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fsrcc"] Nov 24 10:16:41 crc kubenswrapper[4944]: W1124 10:16:41.520755 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0ef4af1_b83d_4edf_99e1_58ae502aa301.slice/crio-45ff467cb212b69c70cbab73e30c5cf27a9940be56f5e442d5701c2fbb892bf4 WatchSource:0}: Error finding container 45ff467cb212b69c70cbab73e30c5cf27a9940be56f5e442d5701c2fbb892bf4: Status 404 returned error can't find the container with id 45ff467cb212b69c70cbab73e30c5cf27a9940be56f5e442d5701c2fbb892bf4 Nov 24 10:16:42 crc kubenswrapper[4944]: I1124 10:16:42.443114 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fsrcc" event={"ID":"d0ef4af1-b83d-4edf-99e1-58ae502aa301","Type":"ContainerStarted","Data":"cd8c44c5fe857a74c32862982e3de75c9f01328b42d198a0fc18363d5ac5b247"} Nov 24 10:16:42 crc kubenswrapper[4944]: I1124 10:16:42.443519 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fsrcc" event={"ID":"d0ef4af1-b83d-4edf-99e1-58ae502aa301","Type":"ContainerStarted","Data":"45ff467cb212b69c70cbab73e30c5cf27a9940be56f5e442d5701c2fbb892bf4"} Nov 24 10:16:42 crc kubenswrapper[4944]: I1124 10:16:42.445642 4944 generic.go:334] "Generic (PLEG): container finished" podID="2ee27e85-9fdc-4440-9747-ae08bf089bd3" containerID="1926e157411d799359d007fd7f7dbb8a420198dc2420df5f96948bf69c65375a" exitCode=0 Nov 24 10:16:42 crc kubenswrapper[4944]: I1124 10:16:42.445684 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" event={"ID":"2ee27e85-9fdc-4440-9747-ae08bf089bd3","Type":"ContainerDied","Data":"1926e157411d799359d007fd7f7dbb8a420198dc2420df5f96948bf69c65375a"} Nov 24 10:16:42 crc kubenswrapper[4944]: I1124 10:16:42.478269 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fsrcc" podStartSLOduration=2.478240899 podStartE2EDuration="2.478240899s" podCreationTimestamp="2025-11-24 10:16:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:42.46980522 +0000 UTC m=+5063.004245682" watchObservedRunningTime="2025-11-24 10:16:42.478240899 +0000 UTC m=+5063.012681391" Nov 24 10:16:43 crc kubenswrapper[4944]: I1124 10:16:43.454864 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" event={"ID":"2ee27e85-9fdc-4440-9747-ae08bf089bd3","Type":"ContainerStarted","Data":"e6d363932a68aad1f31a97717e858580fa6f69acc546642bb868e431dba352ab"} Nov 24 10:16:43 crc kubenswrapper[4944]: I1124 10:16:43.489956 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" podStartSLOduration=3.489926402 podStartE2EDuration="3.489926402s" podCreationTimestamp="2025-11-24 10:16:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:43.486979558 +0000 UTC m=+5064.021420110" watchObservedRunningTime="2025-11-24 10:16:43.489926402 +0000 UTC m=+5064.024366874" Nov 24 10:16:44 crc kubenswrapper[4944]: I1124 10:16:44.462559 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:45 crc kubenswrapper[4944]: I1124 10:16:45.471491 4944 generic.go:334] "Generic (PLEG): container finished" podID="d0ef4af1-b83d-4edf-99e1-58ae502aa301" containerID="cd8c44c5fe857a74c32862982e3de75c9f01328b42d198a0fc18363d5ac5b247" exitCode=0 Nov 24 10:16:45 crc kubenswrapper[4944]: I1124 10:16:45.471577 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fsrcc" event={"ID":"d0ef4af1-b83d-4edf-99e1-58ae502aa301","Type":"ContainerDied","Data":"cd8c44c5fe857a74c32862982e3de75c9f01328b42d198a0fc18363d5ac5b247"} Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.276390 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:16:46 crc kubenswrapper[4944]: E1124 10:16:46.276657 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.796239 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.932870 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rbr2\" (UniqueName: \"kubernetes.io/projected/d0ef4af1-b83d-4edf-99e1-58ae502aa301-kube-api-access-5rbr2\") pod \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.932923 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-credential-keys\") pod \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.932943 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-combined-ca-bundle\") pod \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.933001 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-scripts\") pod \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.933019 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-fernet-keys\") pod \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.933073 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-config-data\") pod \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\" (UID: \"d0ef4af1-b83d-4edf-99e1-58ae502aa301\") " Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.938712 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-scripts" (OuterVolumeSpecName: "scripts") pod "d0ef4af1-b83d-4edf-99e1-58ae502aa301" (UID: "d0ef4af1-b83d-4edf-99e1-58ae502aa301"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.938739 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d0ef4af1-b83d-4edf-99e1-58ae502aa301" (UID: "d0ef4af1-b83d-4edf-99e1-58ae502aa301"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.939097 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0ef4af1-b83d-4edf-99e1-58ae502aa301-kube-api-access-5rbr2" (OuterVolumeSpecName: "kube-api-access-5rbr2") pod "d0ef4af1-b83d-4edf-99e1-58ae502aa301" (UID: "d0ef4af1-b83d-4edf-99e1-58ae502aa301"). InnerVolumeSpecName "kube-api-access-5rbr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.939805 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d0ef4af1-b83d-4edf-99e1-58ae502aa301" (UID: "d0ef4af1-b83d-4edf-99e1-58ae502aa301"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.959794 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-config-data" (OuterVolumeSpecName: "config-data") pod "d0ef4af1-b83d-4edf-99e1-58ae502aa301" (UID: "d0ef4af1-b83d-4edf-99e1-58ae502aa301"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:46 crc kubenswrapper[4944]: I1124 10:16:46.983661 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0ef4af1-b83d-4edf-99e1-58ae502aa301" (UID: "d0ef4af1-b83d-4edf-99e1-58ae502aa301"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.034381 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rbr2\" (UniqueName: \"kubernetes.io/projected/d0ef4af1-b83d-4edf-99e1-58ae502aa301-kube-api-access-5rbr2\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.034577 4944 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.034632 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.034689 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.034738 4944 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.034820 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0ef4af1-b83d-4edf-99e1-58ae502aa301-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.489544 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fsrcc" event={"ID":"d0ef4af1-b83d-4edf-99e1-58ae502aa301","Type":"ContainerDied","Data":"45ff467cb212b69c70cbab73e30c5cf27a9940be56f5e442d5701c2fbb892bf4"} Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.489852 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45ff467cb212b69c70cbab73e30c5cf27a9940be56f5e442d5701c2fbb892bf4" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.489620 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fsrcc" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.549962 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fsrcc"] Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.555514 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fsrcc"] Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.648631 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-pw5fm"] Nov 24 10:16:47 crc kubenswrapper[4944]: E1124 10:16:47.649012 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ef4af1-b83d-4edf-99e1-58ae502aa301" containerName="keystone-bootstrap" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.649030 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ef4af1-b83d-4edf-99e1-58ae502aa301" containerName="keystone-bootstrap" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.649217 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0ef4af1-b83d-4edf-99e1-58ae502aa301" containerName="keystone-bootstrap" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.649911 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.653939 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.653969 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.654004 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.654473 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxvsz" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.654561 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.660931 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pw5fm"] Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.748473 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q82n8\" (UniqueName: \"kubernetes.io/projected/f550d834-f1ef-41df-92c0-dec056f41d0e-kube-api-access-q82n8\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.748510 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-scripts\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.748530 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-combined-ca-bundle\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.748583 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-fernet-keys\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.748642 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-config-data\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.748671 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-credential-keys\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.850644 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-credential-keys\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.850710 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-scripts\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.850765 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q82n8\" (UniqueName: \"kubernetes.io/projected/f550d834-f1ef-41df-92c0-dec056f41d0e-kube-api-access-q82n8\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.850783 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-combined-ca-bundle\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.850842 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-fernet-keys\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.850914 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-config-data\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.856143 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-credential-keys\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.856955 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-scripts\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.857545 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-combined-ca-bundle\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.857782 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-fernet-keys\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.857831 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-config-data\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.866665 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q82n8\" (UniqueName: \"kubernetes.io/projected/f550d834-f1ef-41df-92c0-dec056f41d0e-kube-api-access-q82n8\") pod \"keystone-bootstrap-pw5fm\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:47 crc kubenswrapper[4944]: I1124 10:16:47.967894 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:48 crc kubenswrapper[4944]: I1124 10:16:48.285151 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0ef4af1-b83d-4edf-99e1-58ae502aa301" path="/var/lib/kubelet/pods/d0ef4af1-b83d-4edf-99e1-58ae502aa301/volumes" Nov 24 10:16:48 crc kubenswrapper[4944]: I1124 10:16:48.438786 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pw5fm"] Nov 24 10:16:48 crc kubenswrapper[4944]: W1124 10:16:48.444750 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf550d834_f1ef_41df_92c0_dec056f41d0e.slice/crio-171f260e3ce143bf1b4423ec9899d54a61e4474700f4844147fdb8ed20720898 WatchSource:0}: Error finding container 171f260e3ce143bf1b4423ec9899d54a61e4474700f4844147fdb8ed20720898: Status 404 returned error can't find the container with id 171f260e3ce143bf1b4423ec9899d54a61e4474700f4844147fdb8ed20720898 Nov 24 10:16:48 crc kubenswrapper[4944]: I1124 10:16:48.499526 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pw5fm" event={"ID":"f550d834-f1ef-41df-92c0-dec056f41d0e","Type":"ContainerStarted","Data":"171f260e3ce143bf1b4423ec9899d54a61e4474700f4844147fdb8ed20720898"} Nov 24 10:16:49 crc kubenswrapper[4944]: I1124 10:16:49.507577 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pw5fm" event={"ID":"f550d834-f1ef-41df-92c0-dec056f41d0e","Type":"ContainerStarted","Data":"26d2d171e762e827ca65a927fec0b123e36bd7469d4fa9ee92228d906ebef7f1"} Nov 24 10:16:49 crc kubenswrapper[4944]: I1124 10:16:49.535598 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-pw5fm" podStartSLOduration=2.535578408 podStartE2EDuration="2.535578408s" podCreationTimestamp="2025-11-24 10:16:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:49.522216552 +0000 UTC m=+5070.056657014" watchObservedRunningTime="2025-11-24 10:16:49.535578408 +0000 UTC m=+5070.070018870" Nov 24 10:16:50 crc kubenswrapper[4944]: I1124 10:16:50.970063 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.069883 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-678c494fd7-ggrmn"] Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.070160 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" podUID="ff5bbccb-f586-4bf5-9fc8-5579eb548278" containerName="dnsmasq-dns" containerID="cri-o://a4b386881cb0e5796bea4dfee96632381ecf94287bbc539f1958a645e9737fb5" gracePeriod=10 Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.523860 4944 generic.go:334] "Generic (PLEG): container finished" podID="ff5bbccb-f586-4bf5-9fc8-5579eb548278" containerID="a4b386881cb0e5796bea4dfee96632381ecf94287bbc539f1958a645e9737fb5" exitCode=0 Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.523911 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" event={"ID":"ff5bbccb-f586-4bf5-9fc8-5579eb548278","Type":"ContainerDied","Data":"a4b386881cb0e5796bea4dfee96632381ecf94287bbc539f1958a645e9737fb5"} Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.524291 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" event={"ID":"ff5bbccb-f586-4bf5-9fc8-5579eb548278","Type":"ContainerDied","Data":"6fca2a2237c97de735bf529791226e74fc148d2f3253659f3e9391c89aaa754a"} Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.524308 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fca2a2237c97de735bf529791226e74fc148d2f3253659f3e9391c89aaa754a" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.526678 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.526953 4944 generic.go:334] "Generic (PLEG): container finished" podID="f550d834-f1ef-41df-92c0-dec056f41d0e" containerID="26d2d171e762e827ca65a927fec0b123e36bd7469d4fa9ee92228d906ebef7f1" exitCode=0 Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.527007 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pw5fm" event={"ID":"f550d834-f1ef-41df-92c0-dec056f41d0e","Type":"ContainerDied","Data":"26d2d171e762e827ca65a927fec0b123e36bd7469d4fa9ee92228d906ebef7f1"} Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.605550 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-config\") pod \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.605621 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-sb\") pod \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.605686 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-nb\") pod \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.605789 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-dns-svc\") pod \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.605836 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnjm7\" (UniqueName: \"kubernetes.io/projected/ff5bbccb-f586-4bf5-9fc8-5579eb548278-kube-api-access-wnjm7\") pod \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\" (UID: \"ff5bbccb-f586-4bf5-9fc8-5579eb548278\") " Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.611233 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5bbccb-f586-4bf5-9fc8-5579eb548278-kube-api-access-wnjm7" (OuterVolumeSpecName: "kube-api-access-wnjm7") pod "ff5bbccb-f586-4bf5-9fc8-5579eb548278" (UID: "ff5bbccb-f586-4bf5-9fc8-5579eb548278"). InnerVolumeSpecName "kube-api-access-wnjm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.650125 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-config" (OuterVolumeSpecName: "config") pod "ff5bbccb-f586-4bf5-9fc8-5579eb548278" (UID: "ff5bbccb-f586-4bf5-9fc8-5579eb548278"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.652598 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff5bbccb-f586-4bf5-9fc8-5579eb548278" (UID: "ff5bbccb-f586-4bf5-9fc8-5579eb548278"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.656849 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ff5bbccb-f586-4bf5-9fc8-5579eb548278" (UID: "ff5bbccb-f586-4bf5-9fc8-5579eb548278"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.668608 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff5bbccb-f586-4bf5-9fc8-5579eb548278" (UID: "ff5bbccb-f586-4bf5-9fc8-5579eb548278"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.707811 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.707869 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnjm7\" (UniqueName: \"kubernetes.io/projected/ff5bbccb-f586-4bf5-9fc8-5579eb548278-kube-api-access-wnjm7\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.707880 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.707890 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:51 crc kubenswrapper[4944]: I1124 10:16:51.707898 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff5bbccb-f586-4bf5-9fc8-5579eb548278-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.535133 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-678c494fd7-ggrmn" Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.572492 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-678c494fd7-ggrmn"] Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.587626 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-678c494fd7-ggrmn"] Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.863470 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.930158 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-fernet-keys\") pod \"f550d834-f1ef-41df-92c0-dec056f41d0e\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.930256 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-credential-keys\") pod \"f550d834-f1ef-41df-92c0-dec056f41d0e\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.930322 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-config-data\") pod \"f550d834-f1ef-41df-92c0-dec056f41d0e\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.930445 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q82n8\" (UniqueName: \"kubernetes.io/projected/f550d834-f1ef-41df-92c0-dec056f41d0e-kube-api-access-q82n8\") pod \"f550d834-f1ef-41df-92c0-dec056f41d0e\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.930501 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-combined-ca-bundle\") pod \"f550d834-f1ef-41df-92c0-dec056f41d0e\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.930581 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-scripts\") pod \"f550d834-f1ef-41df-92c0-dec056f41d0e\" (UID: \"f550d834-f1ef-41df-92c0-dec056f41d0e\") " Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.935520 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f550d834-f1ef-41df-92c0-dec056f41d0e-kube-api-access-q82n8" (OuterVolumeSpecName: "kube-api-access-q82n8") pod "f550d834-f1ef-41df-92c0-dec056f41d0e" (UID: "f550d834-f1ef-41df-92c0-dec056f41d0e"). InnerVolumeSpecName "kube-api-access-q82n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.935944 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f550d834-f1ef-41df-92c0-dec056f41d0e" (UID: "f550d834-f1ef-41df-92c0-dec056f41d0e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.936714 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-scripts" (OuterVolumeSpecName: "scripts") pod "f550d834-f1ef-41df-92c0-dec056f41d0e" (UID: "f550d834-f1ef-41df-92c0-dec056f41d0e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.938123 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f550d834-f1ef-41df-92c0-dec056f41d0e" (UID: "f550d834-f1ef-41df-92c0-dec056f41d0e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.957281 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f550d834-f1ef-41df-92c0-dec056f41d0e" (UID: "f550d834-f1ef-41df-92c0-dec056f41d0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:52 crc kubenswrapper[4944]: I1124 10:16:52.974219 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-config-data" (OuterVolumeSpecName: "config-data") pod "f550d834-f1ef-41df-92c0-dec056f41d0e" (UID: "f550d834-f1ef-41df-92c0-dec056f41d0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.032153 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.032208 4944 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.032222 4944 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.032235 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.032247 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q82n8\" (UniqueName: \"kubernetes.io/projected/f550d834-f1ef-41df-92c0-dec056f41d0e-kube-api-access-q82n8\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.032256 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f550d834-f1ef-41df-92c0-dec056f41d0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.543919 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pw5fm" event={"ID":"f550d834-f1ef-41df-92c0-dec056f41d0e","Type":"ContainerDied","Data":"171f260e3ce143bf1b4423ec9899d54a61e4474700f4844147fdb8ed20720898"} Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.543965 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="171f260e3ce143bf1b4423ec9899d54a61e4474700f4844147fdb8ed20720898" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.543993 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pw5fm" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.637591 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8944d6bb7-str5n"] Nov 24 10:16:53 crc kubenswrapper[4944]: E1124 10:16:53.637969 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5bbccb-f586-4bf5-9fc8-5579eb548278" containerName="dnsmasq-dns" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.637989 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5bbccb-f586-4bf5-9fc8-5579eb548278" containerName="dnsmasq-dns" Nov 24 10:16:53 crc kubenswrapper[4944]: E1124 10:16:53.638038 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5bbccb-f586-4bf5-9fc8-5579eb548278" containerName="init" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.638098 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5bbccb-f586-4bf5-9fc8-5579eb548278" containerName="init" Nov 24 10:16:53 crc kubenswrapper[4944]: E1124 10:16:53.638129 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f550d834-f1ef-41df-92c0-dec056f41d0e" containerName="keystone-bootstrap" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.638141 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f550d834-f1ef-41df-92c0-dec056f41d0e" containerName="keystone-bootstrap" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.638343 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5bbccb-f586-4bf5-9fc8-5579eb548278" containerName="dnsmasq-dns" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.638367 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f550d834-f1ef-41df-92c0-dec056f41d0e" containerName="keystone-bootstrap" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.639119 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.642638 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xxvsz" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.642679 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.642638 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.642747 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.653500 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8944d6bb7-str5n"] Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.741328 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-fernet-keys\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.741394 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldn56\" (UniqueName: \"kubernetes.io/projected/aa76b57b-2f12-4a98-a477-25b528e65e55-kube-api-access-ldn56\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.741472 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-scripts\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.741498 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-config-data\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.741550 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-credential-keys\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.741668 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-combined-ca-bundle\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.843553 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-credential-keys\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.843654 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-combined-ca-bundle\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.843722 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-fernet-keys\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.843761 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldn56\" (UniqueName: \"kubernetes.io/projected/aa76b57b-2f12-4a98-a477-25b528e65e55-kube-api-access-ldn56\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.843799 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-scripts\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.843822 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-config-data\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.847221 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-credential-keys\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.847351 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-fernet-keys\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.847516 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-combined-ca-bundle\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.848101 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-config-data\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.850297 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa76b57b-2f12-4a98-a477-25b528e65e55-scripts\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.864784 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldn56\" (UniqueName: \"kubernetes.io/projected/aa76b57b-2f12-4a98-a477-25b528e65e55-kube-api-access-ldn56\") pod \"keystone-8944d6bb7-str5n\" (UID: \"aa76b57b-2f12-4a98-a477-25b528e65e55\") " pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:53 crc kubenswrapper[4944]: I1124 10:16:53.957648 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:54 crc kubenswrapper[4944]: I1124 10:16:54.286674 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff5bbccb-f586-4bf5-9fc8-5579eb548278" path="/var/lib/kubelet/pods/ff5bbccb-f586-4bf5-9fc8-5579eb548278/volumes" Nov 24 10:16:54 crc kubenswrapper[4944]: I1124 10:16:54.375501 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8944d6bb7-str5n"] Nov 24 10:16:54 crc kubenswrapper[4944]: I1124 10:16:54.551268 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8944d6bb7-str5n" event={"ID":"aa76b57b-2f12-4a98-a477-25b528e65e55","Type":"ContainerStarted","Data":"a7c08217ec3ec0527499263b42957028151236a53dcc840b37ea7f0264628064"} Nov 24 10:16:54 crc kubenswrapper[4944]: I1124 10:16:54.551314 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8944d6bb7-str5n" event={"ID":"aa76b57b-2f12-4a98-a477-25b528e65e55","Type":"ContainerStarted","Data":"b0d4a9fdd4a8ccb092b2920acb28f45df4c39835774c03a9a814adc5165ec7bb"} Nov 24 10:16:54 crc kubenswrapper[4944]: I1124 10:16:54.551411 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:16:54 crc kubenswrapper[4944]: I1124 10:16:54.570674 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-8944d6bb7-str5n" podStartSLOduration=1.570650433 podStartE2EDuration="1.570650433s" podCreationTimestamp="2025-11-24 10:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:16:54.568194075 +0000 UTC m=+5075.102634547" watchObservedRunningTime="2025-11-24 10:16:54.570650433 +0000 UTC m=+5075.105090895" Nov 24 10:16:59 crc kubenswrapper[4944]: I1124 10:16:59.276316 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:16:59 crc kubenswrapper[4944]: E1124 10:16:59.276823 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:17:12 crc kubenswrapper[4944]: I1124 10:17:12.277029 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:17:12 crc kubenswrapper[4944]: E1124 10:17:12.277862 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:17:25 crc kubenswrapper[4944]: I1124 10:17:25.455053 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-8944d6bb7-str5n" Nov 24 10:17:27 crc kubenswrapper[4944]: I1124 10:17:27.276990 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:17:27 crc kubenswrapper[4944]: E1124 10:17:27.277773 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:17:29 crc kubenswrapper[4944]: I1124 10:17:29.974904 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 10:17:29 crc kubenswrapper[4944]: I1124 10:17:29.976306 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:17:29 crc kubenswrapper[4944]: I1124 10:17:29.978622 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 10:17:29 crc kubenswrapper[4944]: I1124 10:17:29.978658 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 10:17:29 crc kubenswrapper[4944]: I1124 10:17:29.979721 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-j98zx" Nov 24 10:17:29 crc kubenswrapper[4944]: I1124 10:17:29.984054 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 10:17:29 crc kubenswrapper[4944]: I1124 10:17:29.994500 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 10:17:29 crc kubenswrapper[4944]: I1124 10:17:29.995907 4944 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dbad4e0b-8c7e-4373-a028-74209100a298\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T10:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T10:17:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T10:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T10:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bh59h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T10:17:29Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" not found" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.001917 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 10:17:30 crc kubenswrapper[4944]: E1124 10:17:30.008580 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-bh59h openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[kube-api-access-bh59h openstack-config openstack-config-secret]: context canceled" pod="openstack/openstackclient" podUID="dbad4e0b-8c7e-4373-a028-74209100a298" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.033457 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.034512 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.037915 4944 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dbad4e0b-8c7e-4373-a028-74209100a298" podUID="9e4a6ace-6f5e-4c0e-94e8-82040b15edad" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.044507 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.110583 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m85wt\" (UniqueName: \"kubernetes.io/projected/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-kube-api-access-m85wt\") pod \"openstackclient\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.110657 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config-secret\") pod \"openstackclient\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.110762 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config\") pod \"openstackclient\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.212278 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m85wt\" (UniqueName: \"kubernetes.io/projected/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-kube-api-access-m85wt\") pod \"openstackclient\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.212340 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config-secret\") pod \"openstackclient\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.212418 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config\") pod \"openstackclient\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.213473 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config\") pod \"openstackclient\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.224652 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config-secret\") pod \"openstackclient\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.229604 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m85wt\" (UniqueName: \"kubernetes.io/projected/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-kube-api-access-m85wt\") pod \"openstackclient\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.285943 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbad4e0b-8c7e-4373-a028-74209100a298" path="/var/lib/kubelet/pods/dbad4e0b-8c7e-4373-a028-74209100a298/volumes" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.371134 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.556097 4944 scope.go:117] "RemoveContainer" containerID="53e7aee10dd9a999650fb5910c0a7ec6104f282349fc6388a42c8d33f6c80f53" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.579371 4944 scope.go:117] "RemoveContainer" containerID="1e9c7814f90a47b09017a40258d37a49b1de78b2c47ee2a620422ed80d0b6610" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.614277 4944 scope.go:117] "RemoveContainer" containerID="d6645eb0c62d5fd436446e5734ef76044e695c72011c8d71b17be3403ce8ec10" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.819748 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.870988 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.870979 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9e4a6ace-6f5e-4c0e-94e8-82040b15edad","Type":"ContainerStarted","Data":"8e74859c1f42404209dfec5ce1725eec63e5dfb28b872d90eb68f0dd506f48eb"} Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.876189 4944 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dbad4e0b-8c7e-4373-a028-74209100a298" podUID="9e4a6ace-6f5e-4c0e-94e8-82040b15edad" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.881158 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:17:30 crc kubenswrapper[4944]: I1124 10:17:30.884746 4944 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dbad4e0b-8c7e-4373-a028-74209100a298" podUID="9e4a6ace-6f5e-4c0e-94e8-82040b15edad" Nov 24 10:17:31 crc kubenswrapper[4944]: I1124 10:17:31.880415 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:17:31 crc kubenswrapper[4944]: I1124 10:17:31.880464 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9e4a6ace-6f5e-4c0e-94e8-82040b15edad","Type":"ContainerStarted","Data":"c5a832c4d86aa8f2148f7f59aa26c773201f153873deedab1329c0e606be14bf"} Nov 24 10:17:31 crc kubenswrapper[4944]: I1124 10:17:31.897285 4944 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dbad4e0b-8c7e-4373-a028-74209100a298" podUID="9e4a6ace-6f5e-4c0e-94e8-82040b15edad" Nov 24 10:17:31 crc kubenswrapper[4944]: I1124 10:17:31.903365 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.9033397939999999 podStartE2EDuration="1.903339794s" podCreationTimestamp="2025-11-24 10:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:17:31.893079418 +0000 UTC m=+5112.427519880" watchObservedRunningTime="2025-11-24 10:17:31.903339794 +0000 UTC m=+5112.437780276" Nov 24 10:17:38 crc kubenswrapper[4944]: I1124 10:17:38.277210 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:17:38 crc kubenswrapper[4944]: E1124 10:17:38.278021 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:17:50 crc kubenswrapper[4944]: I1124 10:17:50.282424 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:17:50 crc kubenswrapper[4944]: E1124 10:17:50.283183 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:18:01 crc kubenswrapper[4944]: I1124 10:18:01.276773 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:18:01 crc kubenswrapper[4944]: E1124 10:18:01.277558 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:18:12 crc kubenswrapper[4944]: I1124 10:18:12.278140 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:18:12 crc kubenswrapper[4944]: E1124 10:18:12.279748 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:18:26 crc kubenswrapper[4944]: I1124 10:18:26.277672 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:18:26 crc kubenswrapper[4944]: E1124 10:18:26.279160 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:18:39 crc kubenswrapper[4944]: I1124 10:18:39.278350 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:18:39 crc kubenswrapper[4944]: E1124 10:18:39.279206 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:18:45 crc kubenswrapper[4944]: I1124 10:18:45.833715 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m84r5"] Nov 24 10:18:45 crc kubenswrapper[4944]: I1124 10:18:45.836121 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:45 crc kubenswrapper[4944]: I1124 10:18:45.855584 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m84r5"] Nov 24 10:18:45 crc kubenswrapper[4944]: I1124 10:18:45.946336 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-catalog-content\") pod \"community-operators-m84r5\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:45 crc kubenswrapper[4944]: I1124 10:18:45.946441 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-utilities\") pod \"community-operators-m84r5\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:45 crc kubenswrapper[4944]: I1124 10:18:45.946490 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8k8c\" (UniqueName: \"kubernetes.io/projected/0dd5abec-f975-417c-aa98-b0cc01808485-kube-api-access-q8k8c\") pod \"community-operators-m84r5\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:46 crc kubenswrapper[4944]: I1124 10:18:46.048007 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-catalog-content\") pod \"community-operators-m84r5\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:46 crc kubenswrapper[4944]: I1124 10:18:46.048090 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-utilities\") pod \"community-operators-m84r5\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:46 crc kubenswrapper[4944]: I1124 10:18:46.048142 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8k8c\" (UniqueName: \"kubernetes.io/projected/0dd5abec-f975-417c-aa98-b0cc01808485-kube-api-access-q8k8c\") pod \"community-operators-m84r5\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:46 crc kubenswrapper[4944]: I1124 10:18:46.048486 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-catalog-content\") pod \"community-operators-m84r5\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:46 crc kubenswrapper[4944]: I1124 10:18:46.048587 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-utilities\") pod \"community-operators-m84r5\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:46 crc kubenswrapper[4944]: I1124 10:18:46.071951 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8k8c\" (UniqueName: \"kubernetes.io/projected/0dd5abec-f975-417c-aa98-b0cc01808485-kube-api-access-q8k8c\") pod \"community-operators-m84r5\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:46 crc kubenswrapper[4944]: I1124 10:18:46.156306 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:46 crc kubenswrapper[4944]: I1124 10:18:46.664781 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m84r5"] Nov 24 10:18:47 crc kubenswrapper[4944]: I1124 10:18:47.500828 4944 generic.go:334] "Generic (PLEG): container finished" podID="0dd5abec-f975-417c-aa98-b0cc01808485" containerID="8b00267fc04ca99faa2edc78a82f2e21445220ede7cc43895190b14c8c3389c9" exitCode=0 Nov 24 10:18:47 crc kubenswrapper[4944]: I1124 10:18:47.500903 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m84r5" event={"ID":"0dd5abec-f975-417c-aa98-b0cc01808485","Type":"ContainerDied","Data":"8b00267fc04ca99faa2edc78a82f2e21445220ede7cc43895190b14c8c3389c9"} Nov 24 10:18:47 crc kubenswrapper[4944]: I1124 10:18:47.501171 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m84r5" event={"ID":"0dd5abec-f975-417c-aa98-b0cc01808485","Type":"ContainerStarted","Data":"eca60d7b361a11c998e045486587d85f6b6d56d079ce84cab06bd4ef58beb388"} Nov 24 10:18:47 crc kubenswrapper[4944]: I1124 10:18:47.503200 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 10:18:48 crc kubenswrapper[4944]: I1124 10:18:48.511365 4944 generic.go:334] "Generic (PLEG): container finished" podID="0dd5abec-f975-417c-aa98-b0cc01808485" containerID="04d44c72eb8a1d4019d23b1e67af09bc9e03ed6c8fbef7dc939e157d845164d1" exitCode=0 Nov 24 10:18:48 crc kubenswrapper[4944]: I1124 10:18:48.511444 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m84r5" event={"ID":"0dd5abec-f975-417c-aa98-b0cc01808485","Type":"ContainerDied","Data":"04d44c72eb8a1d4019d23b1e67af09bc9e03ed6c8fbef7dc939e157d845164d1"} Nov 24 10:18:49 crc kubenswrapper[4944]: I1124 10:18:49.519462 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m84r5" event={"ID":"0dd5abec-f975-417c-aa98-b0cc01808485","Type":"ContainerStarted","Data":"06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e"} Nov 24 10:18:49 crc kubenswrapper[4944]: I1124 10:18:49.544269 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m84r5" podStartSLOduration=3.009973853 podStartE2EDuration="4.544245317s" podCreationTimestamp="2025-11-24 10:18:45 +0000 UTC" firstStartedPulling="2025-11-24 10:18:47.502897747 +0000 UTC m=+5188.037338209" lastFinishedPulling="2025-11-24 10:18:49.037169211 +0000 UTC m=+5189.571609673" observedRunningTime="2025-11-24 10:18:49.538903767 +0000 UTC m=+5190.073344229" watchObservedRunningTime="2025-11-24 10:18:49.544245317 +0000 UTC m=+5190.078685789" Nov 24 10:18:54 crc kubenswrapper[4944]: I1124 10:18:54.276827 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:18:54 crc kubenswrapper[4944]: E1124 10:18:54.277612 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:18:56 crc kubenswrapper[4944]: I1124 10:18:56.157247 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:56 crc kubenswrapper[4944]: I1124 10:18:56.157603 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:56 crc kubenswrapper[4944]: I1124 10:18:56.207276 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:56 crc kubenswrapper[4944]: I1124 10:18:56.611490 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:56 crc kubenswrapper[4944]: I1124 10:18:56.651389 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m84r5"] Nov 24 10:18:58 crc kubenswrapper[4944]: I1124 10:18:58.588273 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m84r5" podUID="0dd5abec-f975-417c-aa98-b0cc01808485" containerName="registry-server" containerID="cri-o://06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e" gracePeriod=2 Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.018167 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.195531 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-catalog-content\") pod \"0dd5abec-f975-417c-aa98-b0cc01808485\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.195835 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-utilities\") pod \"0dd5abec-f975-417c-aa98-b0cc01808485\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.195954 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8k8c\" (UniqueName: \"kubernetes.io/projected/0dd5abec-f975-417c-aa98-b0cc01808485-kube-api-access-q8k8c\") pod \"0dd5abec-f975-417c-aa98-b0cc01808485\" (UID: \"0dd5abec-f975-417c-aa98-b0cc01808485\") " Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.196465 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-utilities" (OuterVolumeSpecName: "utilities") pod "0dd5abec-f975-417c-aa98-b0cc01808485" (UID: "0dd5abec-f975-417c-aa98-b0cc01808485"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.201255 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dd5abec-f975-417c-aa98-b0cc01808485-kube-api-access-q8k8c" (OuterVolumeSpecName: "kube-api-access-q8k8c") pod "0dd5abec-f975-417c-aa98-b0cc01808485" (UID: "0dd5abec-f975-417c-aa98-b0cc01808485"). InnerVolumeSpecName "kube-api-access-q8k8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.263207 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0dd5abec-f975-417c-aa98-b0cc01808485" (UID: "0dd5abec-f975-417c-aa98-b0cc01808485"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.298370 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8k8c\" (UniqueName: \"kubernetes.io/projected/0dd5abec-f975-417c-aa98-b0cc01808485-kube-api-access-q8k8c\") on node \"crc\" DevicePath \"\"" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.298401 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.298413 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dd5abec-f975-417c-aa98-b0cc01808485-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.604911 4944 generic.go:334] "Generic (PLEG): container finished" podID="0dd5abec-f975-417c-aa98-b0cc01808485" containerID="06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e" exitCode=0 Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.604957 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m84r5" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.604975 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m84r5" event={"ID":"0dd5abec-f975-417c-aa98-b0cc01808485","Type":"ContainerDied","Data":"06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e"} Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.606188 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m84r5" event={"ID":"0dd5abec-f975-417c-aa98-b0cc01808485","Type":"ContainerDied","Data":"eca60d7b361a11c998e045486587d85f6b6d56d079ce84cab06bd4ef58beb388"} Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.606210 4944 scope.go:117] "RemoveContainer" containerID="06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.627546 4944 scope.go:117] "RemoveContainer" containerID="04d44c72eb8a1d4019d23b1e67af09bc9e03ed6c8fbef7dc939e157d845164d1" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.640585 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m84r5"] Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.648943 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m84r5"] Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.653870 4944 scope.go:117] "RemoveContainer" containerID="8b00267fc04ca99faa2edc78a82f2e21445220ede7cc43895190b14c8c3389c9" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.689927 4944 scope.go:117] "RemoveContainer" containerID="06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e" Nov 24 10:18:59 crc kubenswrapper[4944]: E1124 10:18:59.690528 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e\": container with ID starting with 06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e not found: ID does not exist" containerID="06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.690569 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e"} err="failed to get container status \"06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e\": rpc error: code = NotFound desc = could not find container \"06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e\": container with ID starting with 06d7011ea0b6086d883c44909fce13f9a36767f1a94362f7a9d31f6008d1458e not found: ID does not exist" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.690597 4944 scope.go:117] "RemoveContainer" containerID="04d44c72eb8a1d4019d23b1e67af09bc9e03ed6c8fbef7dc939e157d845164d1" Nov 24 10:18:59 crc kubenswrapper[4944]: E1124 10:18:59.691012 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04d44c72eb8a1d4019d23b1e67af09bc9e03ed6c8fbef7dc939e157d845164d1\": container with ID starting with 04d44c72eb8a1d4019d23b1e67af09bc9e03ed6c8fbef7dc939e157d845164d1 not found: ID does not exist" containerID="04d44c72eb8a1d4019d23b1e67af09bc9e03ed6c8fbef7dc939e157d845164d1" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.691071 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04d44c72eb8a1d4019d23b1e67af09bc9e03ed6c8fbef7dc939e157d845164d1"} err="failed to get container status \"04d44c72eb8a1d4019d23b1e67af09bc9e03ed6c8fbef7dc939e157d845164d1\": rpc error: code = NotFound desc = could not find container \"04d44c72eb8a1d4019d23b1e67af09bc9e03ed6c8fbef7dc939e157d845164d1\": container with ID starting with 04d44c72eb8a1d4019d23b1e67af09bc9e03ed6c8fbef7dc939e157d845164d1 not found: ID does not exist" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.691110 4944 scope.go:117] "RemoveContainer" containerID="8b00267fc04ca99faa2edc78a82f2e21445220ede7cc43895190b14c8c3389c9" Nov 24 10:18:59 crc kubenswrapper[4944]: E1124 10:18:59.691584 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b00267fc04ca99faa2edc78a82f2e21445220ede7cc43895190b14c8c3389c9\": container with ID starting with 8b00267fc04ca99faa2edc78a82f2e21445220ede7cc43895190b14c8c3389c9 not found: ID does not exist" containerID="8b00267fc04ca99faa2edc78a82f2e21445220ede7cc43895190b14c8c3389c9" Nov 24 10:18:59 crc kubenswrapper[4944]: I1124 10:18:59.691627 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b00267fc04ca99faa2edc78a82f2e21445220ede7cc43895190b14c8c3389c9"} err="failed to get container status \"8b00267fc04ca99faa2edc78a82f2e21445220ede7cc43895190b14c8c3389c9\": rpc error: code = NotFound desc = could not find container \"8b00267fc04ca99faa2edc78a82f2e21445220ede7cc43895190b14c8c3389c9\": container with ID starting with 8b00267fc04ca99faa2edc78a82f2e21445220ede7cc43895190b14c8c3389c9 not found: ID does not exist" Nov 24 10:19:00 crc kubenswrapper[4944]: I1124 10:19:00.285439 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dd5abec-f975-417c-aa98-b0cc01808485" path="/var/lib/kubelet/pods/0dd5abec-f975-417c-aa98-b0cc01808485/volumes" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.714926 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-lnchq"] Nov 24 10:19:04 crc kubenswrapper[4944]: E1124 10:19:04.715756 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd5abec-f975-417c-aa98-b0cc01808485" containerName="extract-utilities" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.715773 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd5abec-f975-417c-aa98-b0cc01808485" containerName="extract-utilities" Nov 24 10:19:04 crc kubenswrapper[4944]: E1124 10:19:04.715795 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd5abec-f975-417c-aa98-b0cc01808485" containerName="registry-server" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.715801 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd5abec-f975-417c-aa98-b0cc01808485" containerName="registry-server" Nov 24 10:19:04 crc kubenswrapper[4944]: E1124 10:19:04.715818 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dd5abec-f975-417c-aa98-b0cc01808485" containerName="extract-content" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.715825 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dd5abec-f975-417c-aa98-b0cc01808485" containerName="extract-content" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.716000 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dd5abec-f975-417c-aa98-b0cc01808485" containerName="registry-server" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.716507 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lnchq" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.724574 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lnchq"] Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.809252 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-24fb-account-create-5lptb"] Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.810828 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24fb-account-create-5lptb" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.816072 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.818042 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-24fb-account-create-5lptb"] Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.885191 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w2rr\" (UniqueName: \"kubernetes.io/projected/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-kube-api-access-2w2rr\") pod \"barbican-db-create-lnchq\" (UID: \"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a\") " pod="openstack/barbican-db-create-lnchq" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.885239 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-operator-scripts\") pod \"barbican-db-create-lnchq\" (UID: \"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a\") " pod="openstack/barbican-db-create-lnchq" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.987347 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w2rr\" (UniqueName: \"kubernetes.io/projected/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-kube-api-access-2w2rr\") pod \"barbican-db-create-lnchq\" (UID: \"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a\") " pod="openstack/barbican-db-create-lnchq" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.987400 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-operator-scripts\") pod \"barbican-db-create-lnchq\" (UID: \"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a\") " pod="openstack/barbican-db-create-lnchq" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.987476 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ff6aaed-4959-4839-a354-26c2056dda5e-operator-scripts\") pod \"barbican-24fb-account-create-5lptb\" (UID: \"3ff6aaed-4959-4839-a354-26c2056dda5e\") " pod="openstack/barbican-24fb-account-create-5lptb" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.987517 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fmt9\" (UniqueName: \"kubernetes.io/projected/3ff6aaed-4959-4839-a354-26c2056dda5e-kube-api-access-4fmt9\") pod \"barbican-24fb-account-create-5lptb\" (UID: \"3ff6aaed-4959-4839-a354-26c2056dda5e\") " pod="openstack/barbican-24fb-account-create-5lptb" Nov 24 10:19:04 crc kubenswrapper[4944]: I1124 10:19:04.988446 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-operator-scripts\") pod \"barbican-db-create-lnchq\" (UID: \"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a\") " pod="openstack/barbican-db-create-lnchq" Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.006802 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w2rr\" (UniqueName: \"kubernetes.io/projected/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-kube-api-access-2w2rr\") pod \"barbican-db-create-lnchq\" (UID: \"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a\") " pod="openstack/barbican-db-create-lnchq" Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.049347 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lnchq" Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.089610 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ff6aaed-4959-4839-a354-26c2056dda5e-operator-scripts\") pod \"barbican-24fb-account-create-5lptb\" (UID: \"3ff6aaed-4959-4839-a354-26c2056dda5e\") " pod="openstack/barbican-24fb-account-create-5lptb" Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.089687 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fmt9\" (UniqueName: \"kubernetes.io/projected/3ff6aaed-4959-4839-a354-26c2056dda5e-kube-api-access-4fmt9\") pod \"barbican-24fb-account-create-5lptb\" (UID: \"3ff6aaed-4959-4839-a354-26c2056dda5e\") " pod="openstack/barbican-24fb-account-create-5lptb" Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.091275 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ff6aaed-4959-4839-a354-26c2056dda5e-operator-scripts\") pod \"barbican-24fb-account-create-5lptb\" (UID: \"3ff6aaed-4959-4839-a354-26c2056dda5e\") " pod="openstack/barbican-24fb-account-create-5lptb" Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.105433 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fmt9\" (UniqueName: \"kubernetes.io/projected/3ff6aaed-4959-4839-a354-26c2056dda5e-kube-api-access-4fmt9\") pod \"barbican-24fb-account-create-5lptb\" (UID: \"3ff6aaed-4959-4839-a354-26c2056dda5e\") " pod="openstack/barbican-24fb-account-create-5lptb" Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.134826 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24fb-account-create-5lptb" Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.277552 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:19:05 crc kubenswrapper[4944]: E1124 10:19:05.278101 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.538223 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lnchq"] Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.585717 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-24fb-account-create-5lptb"] Nov 24 10:19:05 crc kubenswrapper[4944]: W1124 10:19:05.586924 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ff6aaed_4959_4839_a354_26c2056dda5e.slice/crio-4098bda10ea2ccc5f104abe2e1a975d798f4a7bc499d5c9b49a0a783c78d5e59 WatchSource:0}: Error finding container 4098bda10ea2ccc5f104abe2e1a975d798f4a7bc499d5c9b49a0a783c78d5e59: Status 404 returned error can't find the container with id 4098bda10ea2ccc5f104abe2e1a975d798f4a7bc499d5c9b49a0a783c78d5e59 Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.650017 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-24fb-account-create-5lptb" event={"ID":"3ff6aaed-4959-4839-a354-26c2056dda5e","Type":"ContainerStarted","Data":"4098bda10ea2ccc5f104abe2e1a975d798f4a7bc499d5c9b49a0a783c78d5e59"} Nov 24 10:19:05 crc kubenswrapper[4944]: I1124 10:19:05.652083 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lnchq" event={"ID":"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a","Type":"ContainerStarted","Data":"4d3666a6a4fdcc0b44052b767b43d9c9a6a37bb027aece795aa11b12eb56b4c4"} Nov 24 10:19:06 crc kubenswrapper[4944]: I1124 10:19:06.659981 4944 generic.go:334] "Generic (PLEG): container finished" podID="d8ee2a60-9953-43cd-8c12-fb7c9e36c85a" containerID="d0778072724d69111fcee9fa5119a3aa735f4eed5a48bd890eb400cb5301487f" exitCode=0 Nov 24 10:19:06 crc kubenswrapper[4944]: I1124 10:19:06.660041 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lnchq" event={"ID":"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a","Type":"ContainerDied","Data":"d0778072724d69111fcee9fa5119a3aa735f4eed5a48bd890eb400cb5301487f"} Nov 24 10:19:06 crc kubenswrapper[4944]: I1124 10:19:06.664213 4944 generic.go:334] "Generic (PLEG): container finished" podID="3ff6aaed-4959-4839-a354-26c2056dda5e" containerID="ac67d78c4a324f1429afe15211ee04394b93fb06e5d16567d56f7f1d01839387" exitCode=0 Nov 24 10:19:06 crc kubenswrapper[4944]: I1124 10:19:06.664249 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-24fb-account-create-5lptb" event={"ID":"3ff6aaed-4959-4839-a354-26c2056dda5e","Type":"ContainerDied","Data":"ac67d78c4a324f1429afe15211ee04394b93fb06e5d16567d56f7f1d01839387"} Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.020999 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lnchq" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.028620 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24fb-account-create-5lptb" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.155937 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w2rr\" (UniqueName: \"kubernetes.io/projected/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-kube-api-access-2w2rr\") pod \"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a\" (UID: \"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a\") " Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.156062 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fmt9\" (UniqueName: \"kubernetes.io/projected/3ff6aaed-4959-4839-a354-26c2056dda5e-kube-api-access-4fmt9\") pod \"3ff6aaed-4959-4839-a354-26c2056dda5e\" (UID: \"3ff6aaed-4959-4839-a354-26c2056dda5e\") " Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.156244 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ff6aaed-4959-4839-a354-26c2056dda5e-operator-scripts\") pod \"3ff6aaed-4959-4839-a354-26c2056dda5e\" (UID: \"3ff6aaed-4959-4839-a354-26c2056dda5e\") " Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.156275 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-operator-scripts\") pod \"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a\" (UID: \"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a\") " Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.156777 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d8ee2a60-9953-43cd-8c12-fb7c9e36c85a" (UID: "d8ee2a60-9953-43cd-8c12-fb7c9e36c85a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.156813 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ff6aaed-4959-4839-a354-26c2056dda5e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3ff6aaed-4959-4839-a354-26c2056dda5e" (UID: "3ff6aaed-4959-4839-a354-26c2056dda5e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.161156 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-kube-api-access-2w2rr" (OuterVolumeSpecName: "kube-api-access-2w2rr") pod "d8ee2a60-9953-43cd-8c12-fb7c9e36c85a" (UID: "d8ee2a60-9953-43cd-8c12-fb7c9e36c85a"). InnerVolumeSpecName "kube-api-access-2w2rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.161241 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ff6aaed-4959-4839-a354-26c2056dda5e-kube-api-access-4fmt9" (OuterVolumeSpecName: "kube-api-access-4fmt9") pod "3ff6aaed-4959-4839-a354-26c2056dda5e" (UID: "3ff6aaed-4959-4839-a354-26c2056dda5e"). InnerVolumeSpecName "kube-api-access-4fmt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.257646 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ff6aaed-4959-4839-a354-26c2056dda5e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.257673 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.257683 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w2rr\" (UniqueName: \"kubernetes.io/projected/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a-kube-api-access-2w2rr\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.257695 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fmt9\" (UniqueName: \"kubernetes.io/projected/3ff6aaed-4959-4839-a354-26c2056dda5e-kube-api-access-4fmt9\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.679802 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-24fb-account-create-5lptb" event={"ID":"3ff6aaed-4959-4839-a354-26c2056dda5e","Type":"ContainerDied","Data":"4098bda10ea2ccc5f104abe2e1a975d798f4a7bc499d5c9b49a0a783c78d5e59"} Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.680182 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4098bda10ea2ccc5f104abe2e1a975d798f4a7bc499d5c9b49a0a783c78d5e59" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.679842 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24fb-account-create-5lptb" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.681509 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lnchq" event={"ID":"d8ee2a60-9953-43cd-8c12-fb7c9e36c85a","Type":"ContainerDied","Data":"4d3666a6a4fdcc0b44052b767b43d9c9a6a37bb027aece795aa11b12eb56b4c4"} Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.681542 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d3666a6a4fdcc0b44052b767b43d9c9a6a37bb027aece795aa11b12eb56b4c4" Nov 24 10:19:08 crc kubenswrapper[4944]: I1124 10:19:08.681631 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lnchq" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.061718 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-qrs6l"] Nov 24 10:19:10 crc kubenswrapper[4944]: E1124 10:19:10.062070 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8ee2a60-9953-43cd-8c12-fb7c9e36c85a" containerName="mariadb-database-create" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.062083 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8ee2a60-9953-43cd-8c12-fb7c9e36c85a" containerName="mariadb-database-create" Nov 24 10:19:10 crc kubenswrapper[4944]: E1124 10:19:10.062119 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ff6aaed-4959-4839-a354-26c2056dda5e" containerName="mariadb-account-create" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.062125 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ff6aaed-4959-4839-a354-26c2056dda5e" containerName="mariadb-account-create" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.062280 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8ee2a60-9953-43cd-8c12-fb7c9e36c85a" containerName="mariadb-database-create" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.062297 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ff6aaed-4959-4839-a354-26c2056dda5e" containerName="mariadb-account-create" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.062861 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.067064 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-flvql" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.067150 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.081033 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qrs6l"] Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.105384 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jvpw\" (UniqueName: \"kubernetes.io/projected/697fd7a2-cd31-45a5-8e12-3b05719a20c0-kube-api-access-8jvpw\") pod \"barbican-db-sync-qrs6l\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.105454 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-db-sync-config-data\") pod \"barbican-db-sync-qrs6l\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.105514 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-combined-ca-bundle\") pod \"barbican-db-sync-qrs6l\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.206516 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-db-sync-config-data\") pod \"barbican-db-sync-qrs6l\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.206597 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-combined-ca-bundle\") pod \"barbican-db-sync-qrs6l\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.206655 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jvpw\" (UniqueName: \"kubernetes.io/projected/697fd7a2-cd31-45a5-8e12-3b05719a20c0-kube-api-access-8jvpw\") pod \"barbican-db-sync-qrs6l\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.212083 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-db-sync-config-data\") pod \"barbican-db-sync-qrs6l\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.212258 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-combined-ca-bundle\") pod \"barbican-db-sync-qrs6l\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.224329 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jvpw\" (UniqueName: \"kubernetes.io/projected/697fd7a2-cd31-45a5-8e12-3b05719a20c0-kube-api-access-8jvpw\") pod \"barbican-db-sync-qrs6l\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.381396 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:10 crc kubenswrapper[4944]: I1124 10:19:10.834337 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qrs6l"] Nov 24 10:19:11 crc kubenswrapper[4944]: I1124 10:19:11.704920 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qrs6l" event={"ID":"697fd7a2-cd31-45a5-8e12-3b05719a20c0","Type":"ContainerStarted","Data":"c1a160deae388ec1c7a13572c6b785e752aa15edebd970415346b103e75f1681"} Nov 24 10:19:11 crc kubenswrapper[4944]: I1124 10:19:11.705260 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qrs6l" event={"ID":"697fd7a2-cd31-45a5-8e12-3b05719a20c0","Type":"ContainerStarted","Data":"48d03ee9a9e3160247d195cfa5b35d16b69b4507e987bfd1fc847410a9a6cfcc"} Nov 24 10:19:11 crc kubenswrapper[4944]: I1124 10:19:11.718861 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-qrs6l" podStartSLOduration=1.718846565 podStartE2EDuration="1.718846565s" podCreationTimestamp="2025-11-24 10:19:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:19:11.716538212 +0000 UTC m=+5212.250978674" watchObservedRunningTime="2025-11-24 10:19:11.718846565 +0000 UTC m=+5212.253287027" Nov 24 10:19:12 crc kubenswrapper[4944]: I1124 10:19:12.714211 4944 generic.go:334] "Generic (PLEG): container finished" podID="697fd7a2-cd31-45a5-8e12-3b05719a20c0" containerID="c1a160deae388ec1c7a13572c6b785e752aa15edebd970415346b103e75f1681" exitCode=0 Nov 24 10:19:12 crc kubenswrapper[4944]: I1124 10:19:12.714316 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qrs6l" event={"ID":"697fd7a2-cd31-45a5-8e12-3b05719a20c0","Type":"ContainerDied","Data":"c1a160deae388ec1c7a13572c6b785e752aa15edebd970415346b103e75f1681"} Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.005979 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.062520 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-combined-ca-bundle\") pod \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.083582 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "697fd7a2-cd31-45a5-8e12-3b05719a20c0" (UID: "697fd7a2-cd31-45a5-8e12-3b05719a20c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.163971 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-db-sync-config-data\") pod \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.164291 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jvpw\" (UniqueName: \"kubernetes.io/projected/697fd7a2-cd31-45a5-8e12-3b05719a20c0-kube-api-access-8jvpw\") pod \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\" (UID: \"697fd7a2-cd31-45a5-8e12-3b05719a20c0\") " Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.164624 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.167178 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "697fd7a2-cd31-45a5-8e12-3b05719a20c0" (UID: "697fd7a2-cd31-45a5-8e12-3b05719a20c0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.167290 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/697fd7a2-cd31-45a5-8e12-3b05719a20c0-kube-api-access-8jvpw" (OuterVolumeSpecName: "kube-api-access-8jvpw") pod "697fd7a2-cd31-45a5-8e12-3b05719a20c0" (UID: "697fd7a2-cd31-45a5-8e12-3b05719a20c0"). InnerVolumeSpecName "kube-api-access-8jvpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.265857 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jvpw\" (UniqueName: \"kubernetes.io/projected/697fd7a2-cd31-45a5-8e12-3b05719a20c0-kube-api-access-8jvpw\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.265889 4944 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/697fd7a2-cd31-45a5-8e12-3b05719a20c0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.732735 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qrs6l" event={"ID":"697fd7a2-cd31-45a5-8e12-3b05719a20c0","Type":"ContainerDied","Data":"48d03ee9a9e3160247d195cfa5b35d16b69b4507e987bfd1fc847410a9a6cfcc"} Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.732793 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48d03ee9a9e3160247d195cfa5b35d16b69b4507e987bfd1fc847410a9a6cfcc" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.732836 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qrs6l" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.909979 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-69767489f-22wtk"] Nov 24 10:19:14 crc kubenswrapper[4944]: E1124 10:19:14.910760 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="697fd7a2-cd31-45a5-8e12-3b05719a20c0" containerName="barbican-db-sync" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.910793 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="697fd7a2-cd31-45a5-8e12-3b05719a20c0" containerName="barbican-db-sync" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.911325 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="697fd7a2-cd31-45a5-8e12-3b05719a20c0" containerName="barbican-db-sync" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.912908 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.917758 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.917883 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.917926 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-flvql" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.927347 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-cfc58c9fb-l9b6d"] Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.928924 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.931175 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.941760 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-69767489f-22wtk"] Nov 24 10:19:14 crc kubenswrapper[4944]: I1124 10:19:14.973914 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-cfc58c9fb-l9b6d"] Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.011511 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf5d847bf-247w5"] Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.013568 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.020092 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf5d847bf-247w5"] Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.079653 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-config-data-custom\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.079915 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dklxb\" (UniqueName: \"kubernetes.io/projected/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-kube-api-access-dklxb\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.080503 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77bc0951-6f9e-4a4f-be92-fadde903228c-logs\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.080641 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-combined-ca-bundle\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.080752 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bc0951-6f9e-4a4f-be92-fadde903228c-combined-ca-bundle\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.080836 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkwvk\" (UniqueName: \"kubernetes.io/projected/77bc0951-6f9e-4a4f-be92-fadde903228c-kube-api-access-dkwvk\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.080918 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-logs\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.080999 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77bc0951-6f9e-4a4f-be92-fadde903228c-config-data-custom\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.081095 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bc0951-6f9e-4a4f-be92-fadde903228c-config-data\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.081254 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-config-data\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.120276 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-668df884fb-fs274"] Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.121973 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.124896 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.146356 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-668df884fb-fs274"] Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.183937 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkwvk\" (UniqueName: \"kubernetes.io/projected/77bc0951-6f9e-4a4f-be92-fadde903228c-kube-api-access-dkwvk\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.184488 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-logs\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.184621 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-nb\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.184694 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77bc0951-6f9e-4a4f-be92-fadde903228c-config-data-custom\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.184779 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bc0951-6f9e-4a4f-be92-fadde903228c-config-data\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.184850 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-dns-svc\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.184967 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-config-data\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.185093 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-config\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.185177 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-sb\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.185288 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz7s6\" (UniqueName: \"kubernetes.io/projected/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-kube-api-access-mz7s6\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.185358 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9836c3a7-f5d2-468c-a3a7-03882013755a-config-data-custom\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.185423 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9836c3a7-f5d2-468c-a3a7-03882013755a-logs\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.185507 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9836c3a7-f5d2-468c-a3a7-03882013755a-config-data\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.185635 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-config-data-custom\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.185740 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjhpg\" (UniqueName: \"kubernetes.io/projected/9836c3a7-f5d2-468c-a3a7-03882013755a-kube-api-access-tjhpg\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.185820 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dklxb\" (UniqueName: \"kubernetes.io/projected/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-kube-api-access-dklxb\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.185924 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77bc0951-6f9e-4a4f-be92-fadde903228c-logs\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.186021 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9836c3a7-f5d2-468c-a3a7-03882013755a-combined-ca-bundle\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.186159 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-combined-ca-bundle\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.186257 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bc0951-6f9e-4a4f-be92-fadde903228c-combined-ca-bundle\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.187281 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77bc0951-6f9e-4a4f-be92-fadde903228c-logs\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.188117 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-logs\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.195627 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77bc0951-6f9e-4a4f-be92-fadde903228c-config-data-custom\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.200370 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77bc0951-6f9e-4a4f-be92-fadde903228c-combined-ca-bundle\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.202813 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dklxb\" (UniqueName: \"kubernetes.io/projected/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-kube-api-access-dklxb\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.203299 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-config-data-custom\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.204599 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-config-data\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.205289 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4-combined-ca-bundle\") pod \"barbican-keystone-listener-cfc58c9fb-l9b6d\" (UID: \"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4\") " pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.205352 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77bc0951-6f9e-4a4f-be92-fadde903228c-config-data\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.214888 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkwvk\" (UniqueName: \"kubernetes.io/projected/77bc0951-6f9e-4a4f-be92-fadde903228c-kube-api-access-dkwvk\") pod \"barbican-worker-69767489f-22wtk\" (UID: \"77bc0951-6f9e-4a4f-be92-fadde903228c\") " pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.275152 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-69767489f-22wtk" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.288538 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-nb\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.288607 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-dns-svc\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.288677 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-config\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.288699 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-sb\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.288785 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz7s6\" (UniqueName: \"kubernetes.io/projected/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-kube-api-access-mz7s6\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.288812 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9836c3a7-f5d2-468c-a3a7-03882013755a-config-data-custom\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.288832 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9836c3a7-f5d2-468c-a3a7-03882013755a-logs\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.288863 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9836c3a7-f5d2-468c-a3a7-03882013755a-config-data\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.288936 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjhpg\" (UniqueName: \"kubernetes.io/projected/9836c3a7-f5d2-468c-a3a7-03882013755a-kube-api-access-tjhpg\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.288999 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9836c3a7-f5d2-468c-a3a7-03882013755a-combined-ca-bundle\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.289497 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9836c3a7-f5d2-468c-a3a7-03882013755a-logs\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.289497 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-dns-svc\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.289563 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-nb\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.289799 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-sb\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.290514 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-config\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.293032 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9836c3a7-f5d2-468c-a3a7-03882013755a-combined-ca-bundle\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.293266 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9836c3a7-f5d2-468c-a3a7-03882013755a-config-data-custom\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.294022 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.294694 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9836c3a7-f5d2-468c-a3a7-03882013755a-config-data\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.316234 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjhpg\" (UniqueName: \"kubernetes.io/projected/9836c3a7-f5d2-468c-a3a7-03882013755a-kube-api-access-tjhpg\") pod \"barbican-api-668df884fb-fs274\" (UID: \"9836c3a7-f5d2-468c-a3a7-03882013755a\") " pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.316614 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz7s6\" (UniqueName: \"kubernetes.io/projected/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-kube-api-access-mz7s6\") pod \"dnsmasq-dns-cf5d847bf-247w5\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.344706 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.446945 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.822959 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf5d847bf-247w5"] Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.831133 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-69767489f-22wtk"] Nov 24 10:19:15 crc kubenswrapper[4944]: I1124 10:19:15.981783 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-cfc58c9fb-l9b6d"] Nov 24 10:19:15 crc kubenswrapper[4944]: W1124 10:19:15.985316 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c84b0f6_464d_4ab3_8fe5_19bb31a2a1b4.slice/crio-be2eb71f60e73397577e9679996957744a1644f9e67f656d29c9252789cc1e51 WatchSource:0}: Error finding container be2eb71f60e73397577e9679996957744a1644f9e67f656d29c9252789cc1e51: Status 404 returned error can't find the container with id be2eb71f60e73397577e9679996957744a1644f9e67f656d29c9252789cc1e51 Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.030090 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-668df884fb-fs274"] Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.276560 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:19:16 crc kubenswrapper[4944]: E1124 10:19:16.277028 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.752008 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69767489f-22wtk" event={"ID":"77bc0951-6f9e-4a4f-be92-fadde903228c","Type":"ContainerStarted","Data":"253106ee51529373adce6f5fb19b89be51c7c581ab46c210092bfacd54fc5e43"} Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.752088 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69767489f-22wtk" event={"ID":"77bc0951-6f9e-4a4f-be92-fadde903228c","Type":"ContainerStarted","Data":"7063d8850e75b2371c43ffe6ff0eb65acb549226f01e46b6efda1fec4cbe6fd4"} Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.752104 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69767489f-22wtk" event={"ID":"77bc0951-6f9e-4a4f-be92-fadde903228c","Type":"ContainerStarted","Data":"6bee22cd098738d6e7ca5d2d18b2c68528ae706dbde5fddb0dcca812ccbc38b5"} Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.754499 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-668df884fb-fs274" event={"ID":"9836c3a7-f5d2-468c-a3a7-03882013755a","Type":"ContainerStarted","Data":"dac071f38d20b3d7cba64df1aac8b1118cee7bf6dddd179eed0c4488da31eb4e"} Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.754545 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-668df884fb-fs274" event={"ID":"9836c3a7-f5d2-468c-a3a7-03882013755a","Type":"ContainerStarted","Data":"53b78ba7dd33db782c8f7aaac413efe4c83a021256764f3b370c71b2ead44764"} Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.754560 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-668df884fb-fs274" event={"ID":"9836c3a7-f5d2-468c-a3a7-03882013755a","Type":"ContainerStarted","Data":"8a7c6137479be5c37793dfce187f9c387b11eec58fd584e2a98b7487cca2b38b"} Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.754604 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.754624 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.756368 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" event={"ID":"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4","Type":"ContainerStarted","Data":"cd30b8cc3a920e049a1f1a402feb8ec797b6dd5802eb5de49c16076cd6b69018"} Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.756410 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" event={"ID":"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4","Type":"ContainerStarted","Data":"028f0c78e73bf00f394b8b183ecbdc35c2f2df850b7555c0c3374d95516854f8"} Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.756423 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" event={"ID":"0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4","Type":"ContainerStarted","Data":"be2eb71f60e73397577e9679996957744a1644f9e67f656d29c9252789cc1e51"} Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.757685 4944 generic.go:334] "Generic (PLEG): container finished" podID="3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" containerID="3d7ded60387711ffceebe376145f6da1928360765b33f0349d43796333ffc7a4" exitCode=0 Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.757733 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" event={"ID":"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef","Type":"ContainerDied","Data":"3d7ded60387711ffceebe376145f6da1928360765b33f0349d43796333ffc7a4"} Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.757853 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" event={"ID":"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef","Type":"ContainerStarted","Data":"cb558ce950fd1bcc3a90eba8b0aea25ba75463080c5a59bcedeb0afe3dd8ff88"} Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.806734 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-69767489f-22wtk" podStartSLOduration=2.806720372 podStartE2EDuration="2.806720372s" podCreationTimestamp="2025-11-24 10:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:19:16.779357971 +0000 UTC m=+5217.313798433" watchObservedRunningTime="2025-11-24 10:19:16.806720372 +0000 UTC m=+5217.341160834" Nov 24 10:19:16 crc kubenswrapper[4944]: I1124 10:19:16.808007 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-668df884fb-fs274" podStartSLOduration=1.808002943 podStartE2EDuration="1.808002943s" podCreationTimestamp="2025-11-24 10:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:19:16.806719152 +0000 UTC m=+5217.341159624" watchObservedRunningTime="2025-11-24 10:19:16.808002943 +0000 UTC m=+5217.342443405" Nov 24 10:19:17 crc kubenswrapper[4944]: I1124 10:19:17.767583 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" event={"ID":"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef","Type":"ContainerStarted","Data":"8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e"} Nov 24 10:19:17 crc kubenswrapper[4944]: I1124 10:19:17.768474 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:17 crc kubenswrapper[4944]: I1124 10:19:17.794657 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-cfc58c9fb-l9b6d" podStartSLOduration=3.794639499 podStartE2EDuration="3.794639499s" podCreationTimestamp="2025-11-24 10:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:19:16.859677138 +0000 UTC m=+5217.394117600" watchObservedRunningTime="2025-11-24 10:19:17.794639499 +0000 UTC m=+5218.329079961" Nov 24 10:19:25 crc kubenswrapper[4944]: I1124 10:19:25.345752 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:19:25 crc kubenswrapper[4944]: I1124 10:19:25.374112 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" podStartSLOduration=11.374087094 podStartE2EDuration="11.374087094s" podCreationTimestamp="2025-11-24 10:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:19:17.804660388 +0000 UTC m=+5218.339100850" watchObservedRunningTime="2025-11-24 10:19:25.374087094 +0000 UTC m=+5225.908527556" Nov 24 10:19:25 crc kubenswrapper[4944]: I1124 10:19:25.416192 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75dd7fd5b7-c8clf"] Nov 24 10:19:25 crc kubenswrapper[4944]: I1124 10:19:25.416473 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" podUID="2ee27e85-9fdc-4440-9747-ae08bf089bd3" containerName="dnsmasq-dns" containerID="cri-o://e6d363932a68aad1f31a97717e858580fa6f69acc546642bb868e431dba352ab" gracePeriod=10 Nov 24 10:19:25 crc kubenswrapper[4944]: I1124 10:19:25.824655 4944 generic.go:334] "Generic (PLEG): container finished" podID="2ee27e85-9fdc-4440-9747-ae08bf089bd3" containerID="e6d363932a68aad1f31a97717e858580fa6f69acc546642bb868e431dba352ab" exitCode=0 Nov 24 10:19:25 crc kubenswrapper[4944]: I1124 10:19:25.824704 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" event={"ID":"2ee27e85-9fdc-4440-9747-ae08bf089bd3","Type":"ContainerDied","Data":"e6d363932a68aad1f31a97717e858580fa6f69acc546642bb868e431dba352ab"} Nov 24 10:19:25 crc kubenswrapper[4944]: I1124 10:19:25.935988 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.066410 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-nb\") pod \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.066778 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rds5\" (UniqueName: \"kubernetes.io/projected/2ee27e85-9fdc-4440-9747-ae08bf089bd3-kube-api-access-2rds5\") pod \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.066812 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-sb\") pod \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.066850 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-config\") pod \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.066953 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-dns-svc\") pod \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\" (UID: \"2ee27e85-9fdc-4440-9747-ae08bf089bd3\") " Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.087616 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ee27e85-9fdc-4440-9747-ae08bf089bd3-kube-api-access-2rds5" (OuterVolumeSpecName: "kube-api-access-2rds5") pod "2ee27e85-9fdc-4440-9747-ae08bf089bd3" (UID: "2ee27e85-9fdc-4440-9747-ae08bf089bd3"). InnerVolumeSpecName "kube-api-access-2rds5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.107151 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2ee27e85-9fdc-4440-9747-ae08bf089bd3" (UID: "2ee27e85-9fdc-4440-9747-ae08bf089bd3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.107615 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-config" (OuterVolumeSpecName: "config") pod "2ee27e85-9fdc-4440-9747-ae08bf089bd3" (UID: "2ee27e85-9fdc-4440-9747-ae08bf089bd3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.126102 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2ee27e85-9fdc-4440-9747-ae08bf089bd3" (UID: "2ee27e85-9fdc-4440-9747-ae08bf089bd3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.138264 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2ee27e85-9fdc-4440-9747-ae08bf089bd3" (UID: "2ee27e85-9fdc-4440-9747-ae08bf089bd3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.182930 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.182958 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.182969 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rds5\" (UniqueName: \"kubernetes.io/projected/2ee27e85-9fdc-4440-9747-ae08bf089bd3-kube-api-access-2rds5\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.182978 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.182986 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ee27e85-9fdc-4440-9747-ae08bf089bd3-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.834389 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" event={"ID":"2ee27e85-9fdc-4440-9747-ae08bf089bd3","Type":"ContainerDied","Data":"beb2e81c0c788a97aee9cab316426bcae4c57190a5c80e6a13f7cbe10abc1c6d"} Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.834445 4944 scope.go:117] "RemoveContainer" containerID="e6d363932a68aad1f31a97717e858580fa6f69acc546642bb868e431dba352ab" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.834562 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dd7fd5b7-c8clf" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.859487 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75dd7fd5b7-c8clf"] Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.859562 4944 scope.go:117] "RemoveContainer" containerID="1926e157411d799359d007fd7f7dbb8a420198dc2420df5f96948bf69c65375a" Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.875701 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75dd7fd5b7-c8clf"] Nov 24 10:19:26 crc kubenswrapper[4944]: I1124 10:19:26.924011 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:27 crc kubenswrapper[4944]: I1124 10:19:27.064770 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-668df884fb-fs274" Nov 24 10:19:28 crc kubenswrapper[4944]: I1124 10:19:28.276989 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:19:28 crc kubenswrapper[4944]: E1124 10:19:28.277506 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:19:28 crc kubenswrapper[4944]: I1124 10:19:28.288386 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ee27e85-9fdc-4440-9747-ae08bf089bd3" path="/var/lib/kubelet/pods/2ee27e85-9fdc-4440-9747-ae08bf089bd3/volumes" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.543458 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-kwp4n"] Nov 24 10:19:38 crc kubenswrapper[4944]: E1124 10:19:38.545516 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee27e85-9fdc-4440-9747-ae08bf089bd3" containerName="dnsmasq-dns" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.545630 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee27e85-9fdc-4440-9747-ae08bf089bd3" containerName="dnsmasq-dns" Nov 24 10:19:38 crc kubenswrapper[4944]: E1124 10:19:38.545733 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee27e85-9fdc-4440-9747-ae08bf089bd3" containerName="init" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.545819 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee27e85-9fdc-4440-9747-ae08bf089bd3" containerName="init" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.546135 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee27e85-9fdc-4440-9747-ae08bf089bd3" containerName="dnsmasq-dns" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.565616 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-kwp4n"] Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.565777 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kwp4n" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.640233 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-071a-account-create-b86qt"] Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.641647 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-071a-account-create-b86qt" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.644271 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.646146 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-071a-account-create-b86qt"] Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.687145 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjx4n\" (UniqueName: \"kubernetes.io/projected/123e08ab-6f70-4f18-a56b-149b8879fd0c-kube-api-access-wjx4n\") pod \"neutron-db-create-kwp4n\" (UID: \"123e08ab-6f70-4f18-a56b-149b8879fd0c\") " pod="openstack/neutron-db-create-kwp4n" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.687309 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/123e08ab-6f70-4f18-a56b-149b8879fd0c-operator-scripts\") pod \"neutron-db-create-kwp4n\" (UID: \"123e08ab-6f70-4f18-a56b-149b8879fd0c\") " pod="openstack/neutron-db-create-kwp4n" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.789300 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf2mr\" (UniqueName: \"kubernetes.io/projected/f2da0f3e-f399-432d-81c3-6a20b7483511-kube-api-access-cf2mr\") pod \"neutron-071a-account-create-b86qt\" (UID: \"f2da0f3e-f399-432d-81c3-6a20b7483511\") " pod="openstack/neutron-071a-account-create-b86qt" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.789467 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjx4n\" (UniqueName: \"kubernetes.io/projected/123e08ab-6f70-4f18-a56b-149b8879fd0c-kube-api-access-wjx4n\") pod \"neutron-db-create-kwp4n\" (UID: \"123e08ab-6f70-4f18-a56b-149b8879fd0c\") " pod="openstack/neutron-db-create-kwp4n" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.789538 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2da0f3e-f399-432d-81c3-6a20b7483511-operator-scripts\") pod \"neutron-071a-account-create-b86qt\" (UID: \"f2da0f3e-f399-432d-81c3-6a20b7483511\") " pod="openstack/neutron-071a-account-create-b86qt" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.789573 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/123e08ab-6f70-4f18-a56b-149b8879fd0c-operator-scripts\") pod \"neutron-db-create-kwp4n\" (UID: \"123e08ab-6f70-4f18-a56b-149b8879fd0c\") " pod="openstack/neutron-db-create-kwp4n" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.790213 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/123e08ab-6f70-4f18-a56b-149b8879fd0c-operator-scripts\") pod \"neutron-db-create-kwp4n\" (UID: \"123e08ab-6f70-4f18-a56b-149b8879fd0c\") " pod="openstack/neutron-db-create-kwp4n" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.809838 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjx4n\" (UniqueName: \"kubernetes.io/projected/123e08ab-6f70-4f18-a56b-149b8879fd0c-kube-api-access-wjx4n\") pod \"neutron-db-create-kwp4n\" (UID: \"123e08ab-6f70-4f18-a56b-149b8879fd0c\") " pod="openstack/neutron-db-create-kwp4n" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.884646 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kwp4n" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.896734 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2da0f3e-f399-432d-81c3-6a20b7483511-operator-scripts\") pod \"neutron-071a-account-create-b86qt\" (UID: \"f2da0f3e-f399-432d-81c3-6a20b7483511\") " pod="openstack/neutron-071a-account-create-b86qt" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.896817 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf2mr\" (UniqueName: \"kubernetes.io/projected/f2da0f3e-f399-432d-81c3-6a20b7483511-kube-api-access-cf2mr\") pod \"neutron-071a-account-create-b86qt\" (UID: \"f2da0f3e-f399-432d-81c3-6a20b7483511\") " pod="openstack/neutron-071a-account-create-b86qt" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.897850 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2da0f3e-f399-432d-81c3-6a20b7483511-operator-scripts\") pod \"neutron-071a-account-create-b86qt\" (UID: \"f2da0f3e-f399-432d-81c3-6a20b7483511\") " pod="openstack/neutron-071a-account-create-b86qt" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.913590 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf2mr\" (UniqueName: \"kubernetes.io/projected/f2da0f3e-f399-432d-81c3-6a20b7483511-kube-api-access-cf2mr\") pod \"neutron-071a-account-create-b86qt\" (UID: \"f2da0f3e-f399-432d-81c3-6a20b7483511\") " pod="openstack/neutron-071a-account-create-b86qt" Nov 24 10:19:38 crc kubenswrapper[4944]: I1124 10:19:38.959227 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-071a-account-create-b86qt" Nov 24 10:19:39 crc kubenswrapper[4944]: I1124 10:19:39.298354 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-kwp4n"] Nov 24 10:19:39 crc kubenswrapper[4944]: W1124 10:19:39.302566 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod123e08ab_6f70_4f18_a56b_149b8879fd0c.slice/crio-4130780fab7bf9d327052a68ce52e259653f031387b22b2797dea6b3c7e9dca8 WatchSource:0}: Error finding container 4130780fab7bf9d327052a68ce52e259653f031387b22b2797dea6b3c7e9dca8: Status 404 returned error can't find the container with id 4130780fab7bf9d327052a68ce52e259653f031387b22b2797dea6b3c7e9dca8 Nov 24 10:19:39 crc kubenswrapper[4944]: I1124 10:19:39.392407 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-071a-account-create-b86qt"] Nov 24 10:19:39 crc kubenswrapper[4944]: W1124 10:19:39.400719 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2da0f3e_f399_432d_81c3_6a20b7483511.slice/crio-c64ba005c053ba0a1d38303478655dca2ffb1c1301b2a5793eae2d00d095208c WatchSource:0}: Error finding container c64ba005c053ba0a1d38303478655dca2ffb1c1301b2a5793eae2d00d095208c: Status 404 returned error can't find the container with id c64ba005c053ba0a1d38303478655dca2ffb1c1301b2a5793eae2d00d095208c Nov 24 10:19:39 crc kubenswrapper[4944]: I1124 10:19:39.926742 4944 generic.go:334] "Generic (PLEG): container finished" podID="123e08ab-6f70-4f18-a56b-149b8879fd0c" containerID="6ffb42869327f7cfcd9e56622a185ddbb74cb32fe121b336bd6221ca41b5f711" exitCode=0 Nov 24 10:19:39 crc kubenswrapper[4944]: I1124 10:19:39.926821 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-kwp4n" event={"ID":"123e08ab-6f70-4f18-a56b-149b8879fd0c","Type":"ContainerDied","Data":"6ffb42869327f7cfcd9e56622a185ddbb74cb32fe121b336bd6221ca41b5f711"} Nov 24 10:19:39 crc kubenswrapper[4944]: I1124 10:19:39.926852 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-kwp4n" event={"ID":"123e08ab-6f70-4f18-a56b-149b8879fd0c","Type":"ContainerStarted","Data":"4130780fab7bf9d327052a68ce52e259653f031387b22b2797dea6b3c7e9dca8"} Nov 24 10:19:39 crc kubenswrapper[4944]: I1124 10:19:39.928031 4944 generic.go:334] "Generic (PLEG): container finished" podID="f2da0f3e-f399-432d-81c3-6a20b7483511" containerID="ec46b567ec5387c476237532f4c35144e9bf880ec8434b125bcedaa596d38894" exitCode=0 Nov 24 10:19:39 crc kubenswrapper[4944]: I1124 10:19:39.928081 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-071a-account-create-b86qt" event={"ID":"f2da0f3e-f399-432d-81c3-6a20b7483511","Type":"ContainerDied","Data":"ec46b567ec5387c476237532f4c35144e9bf880ec8434b125bcedaa596d38894"} Nov 24 10:19:39 crc kubenswrapper[4944]: I1124 10:19:39.928099 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-071a-account-create-b86qt" event={"ID":"f2da0f3e-f399-432d-81c3-6a20b7483511","Type":"ContainerStarted","Data":"c64ba005c053ba0a1d38303478655dca2ffb1c1301b2a5793eae2d00d095208c"} Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.308740 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kwp4n" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.315167 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-071a-account-create-b86qt" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.440843 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2da0f3e-f399-432d-81c3-6a20b7483511-operator-scripts\") pod \"f2da0f3e-f399-432d-81c3-6a20b7483511\" (UID: \"f2da0f3e-f399-432d-81c3-6a20b7483511\") " Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.440902 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjx4n\" (UniqueName: \"kubernetes.io/projected/123e08ab-6f70-4f18-a56b-149b8879fd0c-kube-api-access-wjx4n\") pod \"123e08ab-6f70-4f18-a56b-149b8879fd0c\" (UID: \"123e08ab-6f70-4f18-a56b-149b8879fd0c\") " Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.440927 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf2mr\" (UniqueName: \"kubernetes.io/projected/f2da0f3e-f399-432d-81c3-6a20b7483511-kube-api-access-cf2mr\") pod \"f2da0f3e-f399-432d-81c3-6a20b7483511\" (UID: \"f2da0f3e-f399-432d-81c3-6a20b7483511\") " Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.441016 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/123e08ab-6f70-4f18-a56b-149b8879fd0c-operator-scripts\") pod \"123e08ab-6f70-4f18-a56b-149b8879fd0c\" (UID: \"123e08ab-6f70-4f18-a56b-149b8879fd0c\") " Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.441895 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/123e08ab-6f70-4f18-a56b-149b8879fd0c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "123e08ab-6f70-4f18-a56b-149b8879fd0c" (UID: "123e08ab-6f70-4f18-a56b-149b8879fd0c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.442199 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/123e08ab-6f70-4f18-a56b-149b8879fd0c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.442514 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2da0f3e-f399-432d-81c3-6a20b7483511-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f2da0f3e-f399-432d-81c3-6a20b7483511" (UID: "f2da0f3e-f399-432d-81c3-6a20b7483511"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.447077 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2da0f3e-f399-432d-81c3-6a20b7483511-kube-api-access-cf2mr" (OuterVolumeSpecName: "kube-api-access-cf2mr") pod "f2da0f3e-f399-432d-81c3-6a20b7483511" (UID: "f2da0f3e-f399-432d-81c3-6a20b7483511"). InnerVolumeSpecName "kube-api-access-cf2mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.447193 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/123e08ab-6f70-4f18-a56b-149b8879fd0c-kube-api-access-wjx4n" (OuterVolumeSpecName: "kube-api-access-wjx4n") pod "123e08ab-6f70-4f18-a56b-149b8879fd0c" (UID: "123e08ab-6f70-4f18-a56b-149b8879fd0c"). InnerVolumeSpecName "kube-api-access-wjx4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.544968 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2da0f3e-f399-432d-81c3-6a20b7483511-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.545001 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjx4n\" (UniqueName: \"kubernetes.io/projected/123e08ab-6f70-4f18-a56b-149b8879fd0c-kube-api-access-wjx4n\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.545011 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf2mr\" (UniqueName: \"kubernetes.io/projected/f2da0f3e-f399-432d-81c3-6a20b7483511-kube-api-access-cf2mr\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.944001 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-kwp4n" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.944026 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-kwp4n" event={"ID":"123e08ab-6f70-4f18-a56b-149b8879fd0c","Type":"ContainerDied","Data":"4130780fab7bf9d327052a68ce52e259653f031387b22b2797dea6b3c7e9dca8"} Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.944086 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4130780fab7bf9d327052a68ce52e259653f031387b22b2797dea6b3c7e9dca8" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.945532 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-071a-account-create-b86qt" event={"ID":"f2da0f3e-f399-432d-81c3-6a20b7483511","Type":"ContainerDied","Data":"c64ba005c053ba0a1d38303478655dca2ffb1c1301b2a5793eae2d00d095208c"} Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.945670 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c64ba005c053ba0a1d38303478655dca2ffb1c1301b2a5793eae2d00d095208c" Nov 24 10:19:41 crc kubenswrapper[4944]: I1124 10:19:41.945573 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-071a-account-create-b86qt" Nov 24 10:19:43 crc kubenswrapper[4944]: I1124 10:19:43.276682 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:19:43 crc kubenswrapper[4944]: E1124 10:19:43.278120 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:19:43 crc kubenswrapper[4944]: I1124 10:19:43.897672 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-mnjjq"] Nov 24 10:19:43 crc kubenswrapper[4944]: E1124 10:19:43.897987 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="123e08ab-6f70-4f18-a56b-149b8879fd0c" containerName="mariadb-database-create" Nov 24 10:19:43 crc kubenswrapper[4944]: I1124 10:19:43.898004 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="123e08ab-6f70-4f18-a56b-149b8879fd0c" containerName="mariadb-database-create" Nov 24 10:19:43 crc kubenswrapper[4944]: E1124 10:19:43.898018 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2da0f3e-f399-432d-81c3-6a20b7483511" containerName="mariadb-account-create" Nov 24 10:19:43 crc kubenswrapper[4944]: I1124 10:19:43.898024 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2da0f3e-f399-432d-81c3-6a20b7483511" containerName="mariadb-account-create" Nov 24 10:19:43 crc kubenswrapper[4944]: I1124 10:19:43.898181 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2da0f3e-f399-432d-81c3-6a20b7483511" containerName="mariadb-account-create" Nov 24 10:19:43 crc kubenswrapper[4944]: I1124 10:19:43.898201 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="123e08ab-6f70-4f18-a56b-149b8879fd0c" containerName="mariadb-database-create" Nov 24 10:19:43 crc kubenswrapper[4944]: I1124 10:19:43.898703 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:43 crc kubenswrapper[4944]: I1124 10:19:43.900841 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 10:19:43 crc kubenswrapper[4944]: I1124 10:19:43.900867 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rszjk" Nov 24 10:19:43 crc kubenswrapper[4944]: I1124 10:19:43.901728 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 10:19:43 crc kubenswrapper[4944]: I1124 10:19:43.908666 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mnjjq"] Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.000106 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-config\") pod \"neutron-db-sync-mnjjq\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.000176 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-combined-ca-bundle\") pod \"neutron-db-sync-mnjjq\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.001536 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2dfj\" (UniqueName: \"kubernetes.io/projected/06c01ef8-c21d-47a9-80ef-58ade249d002-kube-api-access-q2dfj\") pod \"neutron-db-sync-mnjjq\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.103005 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-config\") pod \"neutron-db-sync-mnjjq\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.103096 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-combined-ca-bundle\") pod \"neutron-db-sync-mnjjq\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.103198 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2dfj\" (UniqueName: \"kubernetes.io/projected/06c01ef8-c21d-47a9-80ef-58ade249d002-kube-api-access-q2dfj\") pod \"neutron-db-sync-mnjjq\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.108508 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-config\") pod \"neutron-db-sync-mnjjq\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.109108 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-combined-ca-bundle\") pod \"neutron-db-sync-mnjjq\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.128111 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2dfj\" (UniqueName: \"kubernetes.io/projected/06c01ef8-c21d-47a9-80ef-58ade249d002-kube-api-access-q2dfj\") pod \"neutron-db-sync-mnjjq\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.216258 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.653608 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mnjjq"] Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.968159 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mnjjq" event={"ID":"06c01ef8-c21d-47a9-80ef-58ade249d002","Type":"ContainerStarted","Data":"f4b8e477abf074b965c95928aed2ca799c57b9974022dc800dace88373936414"} Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.968478 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mnjjq" event={"ID":"06c01ef8-c21d-47a9-80ef-58ade249d002","Type":"ContainerStarted","Data":"ed4a33e23d7408f134079ddef112b144a2154c1756203d6fab6a359bbf5323bb"} Nov 24 10:19:44 crc kubenswrapper[4944]: I1124 10:19:44.986094 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-mnjjq" podStartSLOduration=1.986069803 podStartE2EDuration="1.986069803s" podCreationTimestamp="2025-11-24 10:19:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:19:44.981396794 +0000 UTC m=+5245.515837276" watchObservedRunningTime="2025-11-24 10:19:44.986069803 +0000 UTC m=+5245.520510275" Nov 24 10:19:49 crc kubenswrapper[4944]: I1124 10:19:49.000845 4944 generic.go:334] "Generic (PLEG): container finished" podID="06c01ef8-c21d-47a9-80ef-58ade249d002" containerID="f4b8e477abf074b965c95928aed2ca799c57b9974022dc800dace88373936414" exitCode=0 Nov 24 10:19:49 crc kubenswrapper[4944]: I1124 10:19:49.000969 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mnjjq" event={"ID":"06c01ef8-c21d-47a9-80ef-58ade249d002","Type":"ContainerDied","Data":"f4b8e477abf074b965c95928aed2ca799c57b9974022dc800dace88373936414"} Nov 24 10:19:50 crc kubenswrapper[4944]: I1124 10:19:50.282532 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:50 crc kubenswrapper[4944]: I1124 10:19:50.406699 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-config\") pod \"06c01ef8-c21d-47a9-80ef-58ade249d002\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " Nov 24 10:19:50 crc kubenswrapper[4944]: I1124 10:19:50.406919 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-combined-ca-bundle\") pod \"06c01ef8-c21d-47a9-80ef-58ade249d002\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " Nov 24 10:19:50 crc kubenswrapper[4944]: I1124 10:19:50.406953 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2dfj\" (UniqueName: \"kubernetes.io/projected/06c01ef8-c21d-47a9-80ef-58ade249d002-kube-api-access-q2dfj\") pod \"06c01ef8-c21d-47a9-80ef-58ade249d002\" (UID: \"06c01ef8-c21d-47a9-80ef-58ade249d002\") " Nov 24 10:19:50 crc kubenswrapper[4944]: I1124 10:19:50.416547 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06c01ef8-c21d-47a9-80ef-58ade249d002-kube-api-access-q2dfj" (OuterVolumeSpecName: "kube-api-access-q2dfj") pod "06c01ef8-c21d-47a9-80ef-58ade249d002" (UID: "06c01ef8-c21d-47a9-80ef-58ade249d002"). InnerVolumeSpecName "kube-api-access-q2dfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:19:50 crc kubenswrapper[4944]: I1124 10:19:50.431835 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-config" (OuterVolumeSpecName: "config") pod "06c01ef8-c21d-47a9-80ef-58ade249d002" (UID: "06c01ef8-c21d-47a9-80ef-58ade249d002"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:19:50 crc kubenswrapper[4944]: I1124 10:19:50.432404 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06c01ef8-c21d-47a9-80ef-58ade249d002" (UID: "06c01ef8-c21d-47a9-80ef-58ade249d002"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:19:50 crc kubenswrapper[4944]: I1124 10:19:50.509563 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:50 crc kubenswrapper[4944]: I1124 10:19:50.509603 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2dfj\" (UniqueName: \"kubernetes.io/projected/06c01ef8-c21d-47a9-80ef-58ade249d002-kube-api-access-q2dfj\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:50 crc kubenswrapper[4944]: I1124 10:19:50.509618 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/06c01ef8-c21d-47a9-80ef-58ade249d002-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.024916 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mnjjq" event={"ID":"06c01ef8-c21d-47a9-80ef-58ade249d002","Type":"ContainerDied","Data":"ed4a33e23d7408f134079ddef112b144a2154c1756203d6fab6a359bbf5323bb"} Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.025208 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed4a33e23d7408f134079ddef112b144a2154c1756203d6fab6a359bbf5323bb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.025150 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mnjjq" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.239314 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fcbf9c88f-tdpxb"] Nov 24 10:19:51 crc kubenswrapper[4944]: E1124 10:19:51.239782 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c01ef8-c21d-47a9-80ef-58ade249d002" containerName="neutron-db-sync" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.239806 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c01ef8-c21d-47a9-80ef-58ade249d002" containerName="neutron-db-sync" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.239993 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="06c01ef8-c21d-47a9-80ef-58ade249d002" containerName="neutron-db-sync" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.241242 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.256642 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fcbf9c88f-tdpxb"] Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.321903 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb2hr\" (UniqueName: \"kubernetes.io/projected/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-kube-api-access-nb2hr\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.322034 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-sb\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.322925 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-dns-svc\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.322969 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-config\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.323001 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-nb\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.438514 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-557c4b8bbf-f4p6l"] Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.441440 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.444012 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-sb\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.448396 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.448821 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-dns-svc\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.448914 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-config\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.448957 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-nb\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.449020 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb2hr\" (UniqueName: \"kubernetes.io/projected/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-kube-api-access-nb2hr\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.449918 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-dns-svc\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.449959 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-config\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.450023 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-nb\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.451769 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.455583 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-557c4b8bbf-f4p6l"] Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.457760 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-sb\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.458373 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rszjk" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.483131 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb2hr\" (UniqueName: \"kubernetes.io/projected/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-kube-api-access-nb2hr\") pod \"dnsmasq-dns-7fcbf9c88f-tdpxb\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.552356 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8xsx\" (UniqueName: \"kubernetes.io/projected/55f037d2-e065-4810-92a4-92ed6753b333-kube-api-access-k8xsx\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.552891 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/55f037d2-e065-4810-92a4-92ed6753b333-config\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.553200 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/55f037d2-e065-4810-92a4-92ed6753b333-httpd-config\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.553250 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f037d2-e065-4810-92a4-92ed6753b333-combined-ca-bundle\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.564920 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.655251 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/55f037d2-e065-4810-92a4-92ed6753b333-httpd-config\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.655295 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f037d2-e065-4810-92a4-92ed6753b333-combined-ca-bundle\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.655441 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8xsx\" (UniqueName: \"kubernetes.io/projected/55f037d2-e065-4810-92a4-92ed6753b333-kube-api-access-k8xsx\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.655485 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/55f037d2-e065-4810-92a4-92ed6753b333-config\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.660569 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/55f037d2-e065-4810-92a4-92ed6753b333-httpd-config\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.660889 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/55f037d2-e065-4810-92a4-92ed6753b333-config\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.666069 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f037d2-e065-4810-92a4-92ed6753b333-combined-ca-bundle\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.679696 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8xsx\" (UniqueName: \"kubernetes.io/projected/55f037d2-e065-4810-92a4-92ed6753b333-kube-api-access-k8xsx\") pod \"neutron-557c4b8bbf-f4p6l\" (UID: \"55f037d2-e065-4810-92a4-92ed6753b333\") " pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:51 crc kubenswrapper[4944]: I1124 10:19:51.784295 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:52 crc kubenswrapper[4944]: I1124 10:19:52.014367 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fcbf9c88f-tdpxb"] Nov 24 10:19:52 crc kubenswrapper[4944]: I1124 10:19:52.035378 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" event={"ID":"2b1c4205-8dbd-422a-b35f-6c1e4a463c21","Type":"ContainerStarted","Data":"6ac3b97fca202eec40e950899a48669af9b3ff1d650c9dc9246b32978e0f1b5b"} Nov 24 10:19:52 crc kubenswrapper[4944]: W1124 10:19:52.306589 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55f037d2_e065_4810_92a4_92ed6753b333.slice/crio-577fec8e5152c8504c7c0ec42638cfff8e40999e9d56ed930fff9316ce585c15 WatchSource:0}: Error finding container 577fec8e5152c8504c7c0ec42638cfff8e40999e9d56ed930fff9316ce585c15: Status 404 returned error can't find the container with id 577fec8e5152c8504c7c0ec42638cfff8e40999e9d56ed930fff9316ce585c15 Nov 24 10:19:52 crc kubenswrapper[4944]: I1124 10:19:52.308245 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-557c4b8bbf-f4p6l"] Nov 24 10:19:53 crc kubenswrapper[4944]: I1124 10:19:53.043677 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557c4b8bbf-f4p6l" event={"ID":"55f037d2-e065-4810-92a4-92ed6753b333","Type":"ContainerStarted","Data":"7b3c8b870fb4c5f5e2f19ccfa7a8c1b7173d71e96139ff3270a45836fa93bfb6"} Nov 24 10:19:53 crc kubenswrapper[4944]: I1124 10:19:53.044018 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:19:53 crc kubenswrapper[4944]: I1124 10:19:53.044036 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557c4b8bbf-f4p6l" event={"ID":"55f037d2-e065-4810-92a4-92ed6753b333","Type":"ContainerStarted","Data":"5e9b07f7c37957e5914b6548c0cfb0e20770ae8e4fbc49ec0f65182b27afa7b7"} Nov 24 10:19:53 crc kubenswrapper[4944]: I1124 10:19:53.044070 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-557c4b8bbf-f4p6l" event={"ID":"55f037d2-e065-4810-92a4-92ed6753b333","Type":"ContainerStarted","Data":"577fec8e5152c8504c7c0ec42638cfff8e40999e9d56ed930fff9316ce585c15"} Nov 24 10:19:53 crc kubenswrapper[4944]: I1124 10:19:53.045090 4944 generic.go:334] "Generic (PLEG): container finished" podID="2b1c4205-8dbd-422a-b35f-6c1e4a463c21" containerID="64dfa47e9a59afaac2acd686c2db629e4a5a14facc81c5b51507600349a02562" exitCode=0 Nov 24 10:19:53 crc kubenswrapper[4944]: I1124 10:19:53.045129 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" event={"ID":"2b1c4205-8dbd-422a-b35f-6c1e4a463c21","Type":"ContainerDied","Data":"64dfa47e9a59afaac2acd686c2db629e4a5a14facc81c5b51507600349a02562"} Nov 24 10:19:53 crc kubenswrapper[4944]: I1124 10:19:53.068541 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-557c4b8bbf-f4p6l" podStartSLOduration=2.068523492 podStartE2EDuration="2.068523492s" podCreationTimestamp="2025-11-24 10:19:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:19:53.061020424 +0000 UTC m=+5253.595460896" watchObservedRunningTime="2025-11-24 10:19:53.068523492 +0000 UTC m=+5253.602963954" Nov 24 10:19:54 crc kubenswrapper[4944]: I1124 10:19:54.055898 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" event={"ID":"2b1c4205-8dbd-422a-b35f-6c1e4a463c21","Type":"ContainerStarted","Data":"a3f25262dc115a792d485a0eb7d390d136ca5a27af9fe6ac160460ef0fea96a9"} Nov 24 10:19:54 crc kubenswrapper[4944]: I1124 10:19:54.056228 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:19:54 crc kubenswrapper[4944]: I1124 10:19:54.075959 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" podStartSLOduration=3.0759424 podStartE2EDuration="3.0759424s" podCreationTimestamp="2025-11-24 10:19:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:19:54.075711053 +0000 UTC m=+5254.610151535" watchObservedRunningTime="2025-11-24 10:19:54.0759424 +0000 UTC m=+5254.610382862" Nov 24 10:19:54 crc kubenswrapper[4944]: I1124 10:19:54.278007 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:19:54 crc kubenswrapper[4944]: E1124 10:19:54.278228 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:20:01 crc kubenswrapper[4944]: I1124 10:20:01.566662 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:20:01 crc kubenswrapper[4944]: I1124 10:20:01.612630 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf5d847bf-247w5"] Nov 24 10:20:01 crc kubenswrapper[4944]: I1124 10:20:01.612956 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" podUID="3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" containerName="dnsmasq-dns" containerID="cri-o://8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e" gracePeriod=10 Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.070613 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.118377 4944 generic.go:334] "Generic (PLEG): container finished" podID="3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" containerID="8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e" exitCode=0 Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.118415 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" event={"ID":"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef","Type":"ContainerDied","Data":"8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e"} Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.118443 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" event={"ID":"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef","Type":"ContainerDied","Data":"cb558ce950fd1bcc3a90eba8b0aea25ba75463080c5a59bcedeb0afe3dd8ff88"} Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.118459 4944 scope.go:117] "RemoveContainer" containerID="8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.118482 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf5d847bf-247w5" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.138981 4944 scope.go:117] "RemoveContainer" containerID="3d7ded60387711ffceebe376145f6da1928360765b33f0349d43796333ffc7a4" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.157066 4944 scope.go:117] "RemoveContainer" containerID="8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e" Nov 24 10:20:02 crc kubenswrapper[4944]: E1124 10:20:02.157499 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e\": container with ID starting with 8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e not found: ID does not exist" containerID="8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.157538 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e"} err="failed to get container status \"8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e\": rpc error: code = NotFound desc = could not find container \"8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e\": container with ID starting with 8b92b04d2952ef3b555083c9d4a9f2eb55003cb9b66278c5a864981e0df9bf6e not found: ID does not exist" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.157565 4944 scope.go:117] "RemoveContainer" containerID="3d7ded60387711ffceebe376145f6da1928360765b33f0349d43796333ffc7a4" Nov 24 10:20:02 crc kubenswrapper[4944]: E1124 10:20:02.158010 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d7ded60387711ffceebe376145f6da1928360765b33f0349d43796333ffc7a4\": container with ID starting with 3d7ded60387711ffceebe376145f6da1928360765b33f0349d43796333ffc7a4 not found: ID does not exist" containerID="3d7ded60387711ffceebe376145f6da1928360765b33f0349d43796333ffc7a4" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.158040 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d7ded60387711ffceebe376145f6da1928360765b33f0349d43796333ffc7a4"} err="failed to get container status \"3d7ded60387711ffceebe376145f6da1928360765b33f0349d43796333ffc7a4\": rpc error: code = NotFound desc = could not find container \"3d7ded60387711ffceebe376145f6da1928360765b33f0349d43796333ffc7a4\": container with ID starting with 3d7ded60387711ffceebe376145f6da1928360765b33f0349d43796333ffc7a4 not found: ID does not exist" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.235453 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-dns-svc\") pod \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.235725 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-config\") pod \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.235943 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-sb\") pod \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.236202 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-nb\") pod \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.236884 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mz7s6\" (UniqueName: \"kubernetes.io/projected/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-kube-api-access-mz7s6\") pod \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\" (UID: \"3dcbd9f2-f81b-454c-8024-a6afc4fde0ef\") " Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.242266 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-kube-api-access-mz7s6" (OuterVolumeSpecName: "kube-api-access-mz7s6") pod "3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" (UID: "3dcbd9f2-f81b-454c-8024-a6afc4fde0ef"). InnerVolumeSpecName "kube-api-access-mz7s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.282246 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" (UID: "3dcbd9f2-f81b-454c-8024-a6afc4fde0ef"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.284638 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-config" (OuterVolumeSpecName: "config") pod "3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" (UID: "3dcbd9f2-f81b-454c-8024-a6afc4fde0ef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.295677 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" (UID: "3dcbd9f2-f81b-454c-8024-a6afc4fde0ef"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.299087 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" (UID: "3dcbd9f2-f81b-454c-8024-a6afc4fde0ef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.339161 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.339373 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.339442 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.339499 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.339555 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mz7s6\" (UniqueName: \"kubernetes.io/projected/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef-kube-api-access-mz7s6\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.453458 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf5d847bf-247w5"] Nov 24 10:20:02 crc kubenswrapper[4944]: I1124 10:20:02.459545 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf5d847bf-247w5"] Nov 24 10:20:04 crc kubenswrapper[4944]: I1124 10:20:04.285067 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" path="/var/lib/kubelet/pods/3dcbd9f2-f81b-454c-8024-a6afc4fde0ef/volumes" Nov 24 10:20:09 crc kubenswrapper[4944]: I1124 10:20:09.276776 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:20:09 crc kubenswrapper[4944]: E1124 10:20:09.277244 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:20:21 crc kubenswrapper[4944]: I1124 10:20:21.795018 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-557c4b8bbf-f4p6l" Nov 24 10:20:24 crc kubenswrapper[4944]: I1124 10:20:24.276498 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:20:25 crc kubenswrapper[4944]: I1124 10:20:25.322307 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"3b50540cdc5646c23244da47ca4ab4ee1e53c396cc0d5f9070264c91666bc8f2"} Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.706964 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-m9kvn"] Nov 24 10:20:28 crc kubenswrapper[4944]: E1124 10:20:28.707992 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" containerName="dnsmasq-dns" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.708009 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" containerName="dnsmasq-dns" Nov 24 10:20:28 crc kubenswrapper[4944]: E1124 10:20:28.708031 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" containerName="init" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.708040 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" containerName="init" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.708286 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dcbd9f2-f81b-454c-8024-a6afc4fde0ef" containerName="dnsmasq-dns" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.709129 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-m9kvn" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.717312 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-m9kvn"] Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.804016 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-2a01-account-create-h4kll"] Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.805418 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2a01-account-create-h4kll" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.807682 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.812338 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2a01-account-create-h4kll"] Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.816596 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4abb1c85-a715-4823-ad1e-0a0fbdbae095-operator-scripts\") pod \"glance-db-create-m9kvn\" (UID: \"4abb1c85-a715-4823-ad1e-0a0fbdbae095\") " pod="openstack/glance-db-create-m9kvn" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.816676 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cd94\" (UniqueName: \"kubernetes.io/projected/4abb1c85-a715-4823-ad1e-0a0fbdbae095-kube-api-access-9cd94\") pod \"glance-db-create-m9kvn\" (UID: \"4abb1c85-a715-4823-ad1e-0a0fbdbae095\") " pod="openstack/glance-db-create-m9kvn" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.918698 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-operator-scripts\") pod \"glance-2a01-account-create-h4kll\" (UID: \"b5d435c0-196f-4a8d-89f7-5e4da0de7eed\") " pod="openstack/glance-2a01-account-create-h4kll" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.918903 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4abb1c85-a715-4823-ad1e-0a0fbdbae095-operator-scripts\") pod \"glance-db-create-m9kvn\" (UID: \"4abb1c85-a715-4823-ad1e-0a0fbdbae095\") " pod="openstack/glance-db-create-m9kvn" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.919072 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cd94\" (UniqueName: \"kubernetes.io/projected/4abb1c85-a715-4823-ad1e-0a0fbdbae095-kube-api-access-9cd94\") pod \"glance-db-create-m9kvn\" (UID: \"4abb1c85-a715-4823-ad1e-0a0fbdbae095\") " pod="openstack/glance-db-create-m9kvn" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.919199 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h2lc\" (UniqueName: \"kubernetes.io/projected/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-kube-api-access-7h2lc\") pod \"glance-2a01-account-create-h4kll\" (UID: \"b5d435c0-196f-4a8d-89f7-5e4da0de7eed\") " pod="openstack/glance-2a01-account-create-h4kll" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.919932 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4abb1c85-a715-4823-ad1e-0a0fbdbae095-operator-scripts\") pod \"glance-db-create-m9kvn\" (UID: \"4abb1c85-a715-4823-ad1e-0a0fbdbae095\") " pod="openstack/glance-db-create-m9kvn" Nov 24 10:20:28 crc kubenswrapper[4944]: I1124 10:20:28.937982 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cd94\" (UniqueName: \"kubernetes.io/projected/4abb1c85-a715-4823-ad1e-0a0fbdbae095-kube-api-access-9cd94\") pod \"glance-db-create-m9kvn\" (UID: \"4abb1c85-a715-4823-ad1e-0a0fbdbae095\") " pod="openstack/glance-db-create-m9kvn" Nov 24 10:20:29 crc kubenswrapper[4944]: I1124 10:20:29.021110 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h2lc\" (UniqueName: \"kubernetes.io/projected/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-kube-api-access-7h2lc\") pod \"glance-2a01-account-create-h4kll\" (UID: \"b5d435c0-196f-4a8d-89f7-5e4da0de7eed\") " pod="openstack/glance-2a01-account-create-h4kll" Nov 24 10:20:29 crc kubenswrapper[4944]: I1124 10:20:29.021184 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-operator-scripts\") pod \"glance-2a01-account-create-h4kll\" (UID: \"b5d435c0-196f-4a8d-89f7-5e4da0de7eed\") " pod="openstack/glance-2a01-account-create-h4kll" Nov 24 10:20:29 crc kubenswrapper[4944]: I1124 10:20:29.021932 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-operator-scripts\") pod \"glance-2a01-account-create-h4kll\" (UID: \"b5d435c0-196f-4a8d-89f7-5e4da0de7eed\") " pod="openstack/glance-2a01-account-create-h4kll" Nov 24 10:20:29 crc kubenswrapper[4944]: I1124 10:20:29.032782 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-m9kvn" Nov 24 10:20:29 crc kubenswrapper[4944]: I1124 10:20:29.052178 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h2lc\" (UniqueName: \"kubernetes.io/projected/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-kube-api-access-7h2lc\") pod \"glance-2a01-account-create-h4kll\" (UID: \"b5d435c0-196f-4a8d-89f7-5e4da0de7eed\") " pod="openstack/glance-2a01-account-create-h4kll" Nov 24 10:20:29 crc kubenswrapper[4944]: I1124 10:20:29.120508 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2a01-account-create-h4kll" Nov 24 10:20:29 crc kubenswrapper[4944]: I1124 10:20:29.510720 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-m9kvn"] Nov 24 10:20:29 crc kubenswrapper[4944]: I1124 10:20:29.600916 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2a01-account-create-h4kll"] Nov 24 10:20:30 crc kubenswrapper[4944]: I1124 10:20:30.362420 4944 generic.go:334] "Generic (PLEG): container finished" podID="b5d435c0-196f-4a8d-89f7-5e4da0de7eed" containerID="4cf81980ba0068728aa8849f6f4ca50c9c51f2a391bd84de93b6abde90eefc60" exitCode=0 Nov 24 10:20:30 crc kubenswrapper[4944]: I1124 10:20:30.362508 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2a01-account-create-h4kll" event={"ID":"b5d435c0-196f-4a8d-89f7-5e4da0de7eed","Type":"ContainerDied","Data":"4cf81980ba0068728aa8849f6f4ca50c9c51f2a391bd84de93b6abde90eefc60"} Nov 24 10:20:30 crc kubenswrapper[4944]: I1124 10:20:30.362572 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2a01-account-create-h4kll" event={"ID":"b5d435c0-196f-4a8d-89f7-5e4da0de7eed","Type":"ContainerStarted","Data":"a93d54f3d41a0734f8e341fd72fdade8a7ff42565d233b1937e26613687bedf3"} Nov 24 10:20:30 crc kubenswrapper[4944]: I1124 10:20:30.364461 4944 generic.go:334] "Generic (PLEG): container finished" podID="4abb1c85-a715-4823-ad1e-0a0fbdbae095" containerID="2f9d6d7b6fdf1e8bd9d917412afa0d1ed283f62c5bb5d21ac5fde4828ada893c" exitCode=0 Nov 24 10:20:30 crc kubenswrapper[4944]: I1124 10:20:30.364515 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-m9kvn" event={"ID":"4abb1c85-a715-4823-ad1e-0a0fbdbae095","Type":"ContainerDied","Data":"2f9d6d7b6fdf1e8bd9d917412afa0d1ed283f62c5bb5d21ac5fde4828ada893c"} Nov 24 10:20:30 crc kubenswrapper[4944]: I1124 10:20:30.364544 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-m9kvn" event={"ID":"4abb1c85-a715-4823-ad1e-0a0fbdbae095","Type":"ContainerStarted","Data":"bc92450902d3e7ca600ab9666f71fb0653bbbbbff76de406b211ad5ec699a12f"} Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.735338 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-m9kvn" Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.741082 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2a01-account-create-h4kll" Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.772120 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-operator-scripts\") pod \"b5d435c0-196f-4a8d-89f7-5e4da0de7eed\" (UID: \"b5d435c0-196f-4a8d-89f7-5e4da0de7eed\") " Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.772243 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h2lc\" (UniqueName: \"kubernetes.io/projected/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-kube-api-access-7h2lc\") pod \"b5d435c0-196f-4a8d-89f7-5e4da0de7eed\" (UID: \"b5d435c0-196f-4a8d-89f7-5e4da0de7eed\") " Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.772277 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4abb1c85-a715-4823-ad1e-0a0fbdbae095-operator-scripts\") pod \"4abb1c85-a715-4823-ad1e-0a0fbdbae095\" (UID: \"4abb1c85-a715-4823-ad1e-0a0fbdbae095\") " Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.772404 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cd94\" (UniqueName: \"kubernetes.io/projected/4abb1c85-a715-4823-ad1e-0a0fbdbae095-kube-api-access-9cd94\") pod \"4abb1c85-a715-4823-ad1e-0a0fbdbae095\" (UID: \"4abb1c85-a715-4823-ad1e-0a0fbdbae095\") " Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.772849 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5d435c0-196f-4a8d-89f7-5e4da0de7eed" (UID: "b5d435c0-196f-4a8d-89f7-5e4da0de7eed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.773410 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4abb1c85-a715-4823-ad1e-0a0fbdbae095-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4abb1c85-a715-4823-ad1e-0a0fbdbae095" (UID: "4abb1c85-a715-4823-ad1e-0a0fbdbae095"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.781966 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-kube-api-access-7h2lc" (OuterVolumeSpecName: "kube-api-access-7h2lc") pod "b5d435c0-196f-4a8d-89f7-5e4da0de7eed" (UID: "b5d435c0-196f-4a8d-89f7-5e4da0de7eed"). InnerVolumeSpecName "kube-api-access-7h2lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.782030 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4abb1c85-a715-4823-ad1e-0a0fbdbae095-kube-api-access-9cd94" (OuterVolumeSpecName: "kube-api-access-9cd94") pod "4abb1c85-a715-4823-ad1e-0a0fbdbae095" (UID: "4abb1c85-a715-4823-ad1e-0a0fbdbae095"). InnerVolumeSpecName "kube-api-access-9cd94". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.874996 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cd94\" (UniqueName: \"kubernetes.io/projected/4abb1c85-a715-4823-ad1e-0a0fbdbae095-kube-api-access-9cd94\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.875033 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.875042 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h2lc\" (UniqueName: \"kubernetes.io/projected/b5d435c0-196f-4a8d-89f7-5e4da0de7eed-kube-api-access-7h2lc\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:31 crc kubenswrapper[4944]: I1124 10:20:31.875063 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4abb1c85-a715-4823-ad1e-0a0fbdbae095-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:32 crc kubenswrapper[4944]: I1124 10:20:32.382232 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-m9kvn" event={"ID":"4abb1c85-a715-4823-ad1e-0a0fbdbae095","Type":"ContainerDied","Data":"bc92450902d3e7ca600ab9666f71fb0653bbbbbff76de406b211ad5ec699a12f"} Nov 24 10:20:32 crc kubenswrapper[4944]: I1124 10:20:32.382620 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc92450902d3e7ca600ab9666f71fb0653bbbbbff76de406b211ad5ec699a12f" Nov 24 10:20:32 crc kubenswrapper[4944]: I1124 10:20:32.382327 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-m9kvn" Nov 24 10:20:32 crc kubenswrapper[4944]: I1124 10:20:32.385002 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2a01-account-create-h4kll" event={"ID":"b5d435c0-196f-4a8d-89f7-5e4da0de7eed","Type":"ContainerDied","Data":"a93d54f3d41a0734f8e341fd72fdade8a7ff42565d233b1937e26613687bedf3"} Nov 24 10:20:32 crc kubenswrapper[4944]: I1124 10:20:32.385027 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a93d54f3d41a0734f8e341fd72fdade8a7ff42565d233b1937e26613687bedf3" Nov 24 10:20:32 crc kubenswrapper[4944]: I1124 10:20:32.385089 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2a01-account-create-h4kll" Nov 24 10:20:33 crc kubenswrapper[4944]: I1124 10:20:33.965134 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-r774x"] Nov 24 10:20:33 crc kubenswrapper[4944]: E1124 10:20:33.965471 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4abb1c85-a715-4823-ad1e-0a0fbdbae095" containerName="mariadb-database-create" Nov 24 10:20:33 crc kubenswrapper[4944]: I1124 10:20:33.965484 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4abb1c85-a715-4823-ad1e-0a0fbdbae095" containerName="mariadb-database-create" Nov 24 10:20:33 crc kubenswrapper[4944]: E1124 10:20:33.965507 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d435c0-196f-4a8d-89f7-5e4da0de7eed" containerName="mariadb-account-create" Nov 24 10:20:33 crc kubenswrapper[4944]: I1124 10:20:33.965514 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d435c0-196f-4a8d-89f7-5e4da0de7eed" containerName="mariadb-account-create" Nov 24 10:20:33 crc kubenswrapper[4944]: I1124 10:20:33.965673 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="4abb1c85-a715-4823-ad1e-0a0fbdbae095" containerName="mariadb-database-create" Nov 24 10:20:33 crc kubenswrapper[4944]: I1124 10:20:33.965684 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5d435c0-196f-4a8d-89f7-5e4da0de7eed" containerName="mariadb-account-create" Nov 24 10:20:33 crc kubenswrapper[4944]: I1124 10:20:33.966292 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r774x" Nov 24 10:20:33 crc kubenswrapper[4944]: I1124 10:20:33.969684 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 10:20:33 crc kubenswrapper[4944]: I1124 10:20:33.970178 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-cgvmr" Nov 24 10:20:33 crc kubenswrapper[4944]: I1124 10:20:33.988197 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-r774x"] Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.015617 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-config-data\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.015683 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-combined-ca-bundle\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.015752 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4j5r\" (UniqueName: \"kubernetes.io/projected/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-kube-api-access-z4j5r\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.015807 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-db-sync-config-data\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.116961 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-db-sync-config-data\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.117071 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-config-data\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.117121 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-combined-ca-bundle\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.117192 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4j5r\" (UniqueName: \"kubernetes.io/projected/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-kube-api-access-z4j5r\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.123092 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-db-sync-config-data\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.123186 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-combined-ca-bundle\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.134309 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-config-data\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.139700 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4j5r\" (UniqueName: \"kubernetes.io/projected/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-kube-api-access-z4j5r\") pod \"glance-db-sync-r774x\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.283303 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r774x" Nov 24 10:20:34 crc kubenswrapper[4944]: I1124 10:20:34.911182 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-r774x"] Nov 24 10:20:35 crc kubenswrapper[4944]: I1124 10:20:35.436690 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r774x" event={"ID":"b2c072d8-6141-44f7-a3cb-59679ce2d5ba","Type":"ContainerStarted","Data":"e626e99a790e19a9baf5df1a8816981a5221dfeb618e3888722e791cdbcb85d5"} Nov 24 10:20:36 crc kubenswrapper[4944]: I1124 10:20:36.448197 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r774x" event={"ID":"b2c072d8-6141-44f7-a3cb-59679ce2d5ba","Type":"ContainerStarted","Data":"fdb8f570d4b4ecc9cc9437e14ad556388e848c47a804bde425a413d1a138d1af"} Nov 24 10:20:36 crc kubenswrapper[4944]: I1124 10:20:36.469850 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-r774x" podStartSLOduration=3.469830875 podStartE2EDuration="3.469830875s" podCreationTimestamp="2025-11-24 10:20:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:20:36.461790199 +0000 UTC m=+5296.996230661" watchObservedRunningTime="2025-11-24 10:20:36.469830875 +0000 UTC m=+5297.004271337" Nov 24 10:20:39 crc kubenswrapper[4944]: I1124 10:20:39.471526 4944 generic.go:334] "Generic (PLEG): container finished" podID="b2c072d8-6141-44f7-a3cb-59679ce2d5ba" containerID="fdb8f570d4b4ecc9cc9437e14ad556388e848c47a804bde425a413d1a138d1af" exitCode=0 Nov 24 10:20:39 crc kubenswrapper[4944]: I1124 10:20:39.471617 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r774x" event={"ID":"b2c072d8-6141-44f7-a3cb-59679ce2d5ba","Type":"ContainerDied","Data":"fdb8f570d4b4ecc9cc9437e14ad556388e848c47a804bde425a413d1a138d1af"} Nov 24 10:20:40 crc kubenswrapper[4944]: I1124 10:20:40.853160 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r774x" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.024338 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-config-data\") pod \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.024504 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-db-sync-config-data\") pod \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.025308 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-combined-ca-bundle\") pod \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.025398 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4j5r\" (UniqueName: \"kubernetes.io/projected/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-kube-api-access-z4j5r\") pod \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\" (UID: \"b2c072d8-6141-44f7-a3cb-59679ce2d5ba\") " Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.030325 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b2c072d8-6141-44f7-a3cb-59679ce2d5ba" (UID: "b2c072d8-6141-44f7-a3cb-59679ce2d5ba"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.044178 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-kube-api-access-z4j5r" (OuterVolumeSpecName: "kube-api-access-z4j5r") pod "b2c072d8-6141-44f7-a3cb-59679ce2d5ba" (UID: "b2c072d8-6141-44f7-a3cb-59679ce2d5ba"). InnerVolumeSpecName "kube-api-access-z4j5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.048641 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2c072d8-6141-44f7-a3cb-59679ce2d5ba" (UID: "b2c072d8-6141-44f7-a3cb-59679ce2d5ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.065669 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-config-data" (OuterVolumeSpecName: "config-data") pod "b2c072d8-6141-44f7-a3cb-59679ce2d5ba" (UID: "b2c072d8-6141-44f7-a3cb-59679ce2d5ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.127387 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.127422 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4j5r\" (UniqueName: \"kubernetes.io/projected/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-kube-api-access-z4j5r\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.127435 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.127445 4944 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b2c072d8-6141-44f7-a3cb-59679ce2d5ba-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.506648 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r774x" event={"ID":"b2c072d8-6141-44f7-a3cb-59679ce2d5ba","Type":"ContainerDied","Data":"e626e99a790e19a9baf5df1a8816981a5221dfeb618e3888722e791cdbcb85d5"} Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.506674 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r774x" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.506694 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e626e99a790e19a9baf5df1a8816981a5221dfeb618e3888722e791cdbcb85d5" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.791907 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:20:41 crc kubenswrapper[4944]: E1124 10:20:41.792705 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2c072d8-6141-44f7-a3cb-59679ce2d5ba" containerName="glance-db-sync" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.792730 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2c072d8-6141-44f7-a3cb-59679ce2d5ba" containerName="glance-db-sync" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.792995 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2c072d8-6141-44f7-a3cb-59679ce2d5ba" containerName="glance-db-sync" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.794180 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 10:20:41 crc kubenswrapper[4944]: W1124 10:20:41.798020 4944 reflector.go:561] object-"openstack"/"glance-scripts": failed to list *v1.Secret: secrets "glance-scripts" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Nov 24 10:20:41 crc kubenswrapper[4944]: E1124 10:20:41.798093 4944 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"glance-scripts\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"glance-scripts\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.798362 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-cgvmr" Nov 24 10:20:41 crc kubenswrapper[4944]: W1124 10:20:41.798543 4944 reflector.go:561] object-"openstack"/"glance-default-external-config-data": failed to list *v1.Secret: secrets "glance-default-external-config-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Nov 24 10:20:41 crc kubenswrapper[4944]: E1124 10:20:41.798567 4944 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"glance-default-external-config-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"glance-default-external-config-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 10:20:41 crc kubenswrapper[4944]: W1124 10:20:41.798615 4944 reflector.go:561] object-"openstack"/"ceph-conf-files": failed to list *v1.Secret: secrets "ceph-conf-files" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Nov 24 10:20:41 crc kubenswrapper[4944]: E1124 10:20:41.798626 4944 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"ceph-conf-files\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ceph-conf-files\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.812826 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.936659 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64c4c7d56f-pzxnn"] Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.938196 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.948429 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-scripts\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.948617 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-ceph\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.948687 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-config-data\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.948725 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.948824 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt66h\" (UniqueName: \"kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-kube-api-access-nt66h\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.948887 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:41 crc kubenswrapper[4944]: I1124 10:20:41.949000 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-logs\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.002777 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64c4c7d56f-pzxnn"] Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.051825 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfxsj\" (UniqueName: \"kubernetes.io/projected/f6aaa08f-11fd-495a-88f4-e1684d0f229d-kube-api-access-hfxsj\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.051873 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-sb\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.051919 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt66h\" (UniqueName: \"kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-kube-api-access-nt66h\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.051978 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.052067 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-logs\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.052104 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-dns-svc\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.052132 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-scripts\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.052185 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-ceph\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.052219 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-config\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.052258 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-config-data\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.052281 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-nb\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.052308 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.052594 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-logs\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.052658 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.071482 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.073430 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt66h\" (UniqueName: \"kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-kube-api-access-nt66h\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.077194 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.083479 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.088211 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.088346 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.167992 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-nb\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.168063 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfxsj\" (UniqueName: \"kubernetes.io/projected/f6aaa08f-11fd-495a-88f4-e1684d0f229d-kube-api-access-hfxsj\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.168089 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-sb\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.169477 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-sb\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.169425 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-dns-svc\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.169527 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-nb\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.169684 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-config\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.170209 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-dns-svc\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.170329 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-config\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.191609 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfxsj\" (UniqueName: \"kubernetes.io/projected/f6aaa08f-11fd-495a-88f4-e1684d0f229d-kube-api-access-hfxsj\") pod \"dnsmasq-dns-64c4c7d56f-pzxnn\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.254705 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.271027 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.271109 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.271137 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.271182 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-logs\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.271213 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.271278 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.271360 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs2lv\" (UniqueName: \"kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-kube-api-access-hs2lv\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.378490 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs2lv\" (UniqueName: \"kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-kube-api-access-hs2lv\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.378849 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.378882 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.378897 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.378930 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-logs\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.378951 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.379007 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.380569 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-logs\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.380924 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.394791 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.398411 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs2lv\" (UniqueName: \"kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-kube-api-access-hs2lv\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.399727 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.727338 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.734839 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.736620 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-ceph\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.796523 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64c4c7d56f-pzxnn"] Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.812499 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.817591 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-scripts\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.824693 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:42 crc kubenswrapper[4944]: I1124 10:20:42.836246 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.024466 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:20:43 crc kubenswrapper[4944]: E1124 10:20:43.025693 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config-data], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-external-api-0" podUID="36c45460-feb1-4333-bd27-37ae7213fc2f" Nov 24 10:20:43 crc kubenswrapper[4944]: E1124 10:20:43.053366 4944 secret.go:188] Couldn't get secret openstack/glance-default-external-config-data: failed to sync secret cache: timed out waiting for the condition Nov 24 10:20:43 crc kubenswrapper[4944]: E1124 10:20:43.053447 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-config-data podName:36c45460-feb1-4333-bd27-37ae7213fc2f nodeName:}" failed. No retries permitted until 2025-11-24 10:20:43.553428689 +0000 UTC m=+5304.087869151 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-config-data") pod "glance-default-external-api-0" (UID: "36c45460-feb1-4333-bd27-37ae7213fc2f") : failed to sync secret cache: timed out waiting for the condition Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.362713 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.363938 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.565652 4944 generic.go:334] "Generic (PLEG): container finished" podID="f6aaa08f-11fd-495a-88f4-e1684d0f229d" containerID="bab45faba9c74c08563981c1a099dbef2267aa6c1c11511dc7878fd1de5763bd" exitCode=0 Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.565837 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" event={"ID":"f6aaa08f-11fd-495a-88f4-e1684d0f229d","Type":"ContainerDied","Data":"bab45faba9c74c08563981c1a099dbef2267aa6c1c11511dc7878fd1de5763bd"} Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.565869 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" event={"ID":"f6aaa08f-11fd-495a-88f4-e1684d0f229d","Type":"ContainerStarted","Data":"0a215c6dafaa81931fc9ffd730566bc28058891ddc269dbe1bfd14489b207a0c"} Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.567699 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.568085 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48e5d2b8-092b-4091-8a21-f5166c29ae9b","Type":"ContainerStarted","Data":"9f9587ac15e41f04786dd5931458c522634e538885895536838cef8716beb2bc"} Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.604781 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-config-data\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.609684 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-config-data\") pod \"glance-default-external-api-0\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.689326 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.807883 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-combined-ca-bundle\") pod \"36c45460-feb1-4333-bd27-37ae7213fc2f\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.808480 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-logs\") pod \"36c45460-feb1-4333-bd27-37ae7213fc2f\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.808519 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-scripts\") pod \"36c45460-feb1-4333-bd27-37ae7213fc2f\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.808613 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-ceph\") pod \"36c45460-feb1-4333-bd27-37ae7213fc2f\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.808843 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-logs" (OuterVolumeSpecName: "logs") pod "36c45460-feb1-4333-bd27-37ae7213fc2f" (UID: "36c45460-feb1-4333-bd27-37ae7213fc2f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.808629 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nt66h\" (UniqueName: \"kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-kube-api-access-nt66h\") pod \"36c45460-feb1-4333-bd27-37ae7213fc2f\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.808993 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-config-data\") pod \"36c45460-feb1-4333-bd27-37ae7213fc2f\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.809071 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-httpd-run\") pod \"36c45460-feb1-4333-bd27-37ae7213fc2f\" (UID: \"36c45460-feb1-4333-bd27-37ae7213fc2f\") " Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.809516 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.809820 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "36c45460-feb1-4333-bd27-37ae7213fc2f" (UID: "36c45460-feb1-4333-bd27-37ae7213fc2f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.812310 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-scripts" (OuterVolumeSpecName: "scripts") pod "36c45460-feb1-4333-bd27-37ae7213fc2f" (UID: "36c45460-feb1-4333-bd27-37ae7213fc2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.812412 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36c45460-feb1-4333-bd27-37ae7213fc2f" (UID: "36c45460-feb1-4333-bd27-37ae7213fc2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.815185 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-config-data" (OuterVolumeSpecName: "config-data") pod "36c45460-feb1-4333-bd27-37ae7213fc2f" (UID: "36c45460-feb1-4333-bd27-37ae7213fc2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.815335 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-kube-api-access-nt66h" (OuterVolumeSpecName: "kube-api-access-nt66h") pod "36c45460-feb1-4333-bd27-37ae7213fc2f" (UID: "36c45460-feb1-4333-bd27-37ae7213fc2f"). InnerVolumeSpecName "kube-api-access-nt66h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.815394 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-ceph" (OuterVolumeSpecName: "ceph") pod "36c45460-feb1-4333-bd27-37ae7213fc2f" (UID: "36c45460-feb1-4333-bd27-37ae7213fc2f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.911179 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.911213 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.911225 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.911233 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nt66h\" (UniqueName: \"kubernetes.io/projected/36c45460-feb1-4333-bd27-37ae7213fc2f-kube-api-access-nt66h\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.911244 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c45460-feb1-4333-bd27-37ae7213fc2f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:43 crc kubenswrapper[4944]: I1124 10:20:43.911252 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36c45460-feb1-4333-bd27-37ae7213fc2f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.575604 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48e5d2b8-092b-4091-8a21-f5166c29ae9b","Type":"ContainerStarted","Data":"a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16"} Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.575867 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48e5d2b8-092b-4091-8a21-f5166c29ae9b","Type":"ContainerStarted","Data":"1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472"} Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.578243 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" event={"ID":"f6aaa08f-11fd-495a-88f4-e1684d0f229d","Type":"ContainerStarted","Data":"decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309"} Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.578261 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.578477 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.599556 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.599533259 podStartE2EDuration="2.599533259s" podCreationTimestamp="2025-11-24 10:20:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:20:44.595705168 +0000 UTC m=+5305.130145650" watchObservedRunningTime="2025-11-24 10:20:44.599533259 +0000 UTC m=+5305.133973721" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.618789 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" podStartSLOduration=3.618772623 podStartE2EDuration="3.618772623s" podCreationTimestamp="2025-11-24 10:20:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:20:44.615800218 +0000 UTC m=+5305.150240690" watchObservedRunningTime="2025-11-24 10:20:44.618772623 +0000 UTC m=+5305.153213095" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.664138 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.670094 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.683440 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.685137 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.686896 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.696958 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.826881 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-scripts\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.827481 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-config-data\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.827549 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.827587 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltztr\" (UniqueName: \"kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-kube-api-access-ltztr\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.827718 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-logs\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.827798 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-ceph\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.827964 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.929243 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltztr\" (UniqueName: \"kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-kube-api-access-ltztr\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.929643 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-logs\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.930177 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-logs\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.930252 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-ceph\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.931084 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.931133 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-scripts\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.931185 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-config-data\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.931247 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.931611 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.937211 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.938021 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-scripts\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.938627 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-ceph\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.949592 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltztr\" (UniqueName: \"kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-kube-api-access-ltztr\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.949634 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-config-data\") pod \"glance-default-external-api-0\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " pod="openstack/glance-default-external-api-0" Nov 24 10:20:44 crc kubenswrapper[4944]: I1124 10:20:44.967441 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:20:45 crc kubenswrapper[4944]: I1124 10:20:45.004117 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 10:20:45 crc kubenswrapper[4944]: I1124 10:20:45.580425 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:20:46 crc kubenswrapper[4944]: I1124 10:20:46.288544 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36c45460-feb1-4333-bd27-37ae7213fc2f" path="/var/lib/kubelet/pods/36c45460-feb1-4333-bd27-37ae7213fc2f/volumes" Nov 24 10:20:46 crc kubenswrapper[4944]: I1124 10:20:46.620849 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"57e038b8-73fd-4a98-8d1d-d06df5909909","Type":"ContainerStarted","Data":"4550c9ae6e4446c0b8c01687c8c450d2082237a87c609f713ab0005d55cd8139"} Nov 24 10:20:46 crc kubenswrapper[4944]: I1124 10:20:46.621124 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="48e5d2b8-092b-4091-8a21-f5166c29ae9b" containerName="glance-httpd" containerID="cri-o://a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16" gracePeriod=30 Nov 24 10:20:46 crc kubenswrapper[4944]: I1124 10:20:46.621211 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"57e038b8-73fd-4a98-8d1d-d06df5909909","Type":"ContainerStarted","Data":"56f50d48225e27d57f356dfdf23f611e6f870edabed7672e2657ecac0f0619d6"} Nov 24 10:20:46 crc kubenswrapper[4944]: I1124 10:20:46.621225 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"57e038b8-73fd-4a98-8d1d-d06df5909909","Type":"ContainerStarted","Data":"d66ffe6d68c243de044f8f352e1b7f1e90fcf90fbb837051f10aa96f82349c51"} Nov 24 10:20:46 crc kubenswrapper[4944]: I1124 10:20:46.620959 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="48e5d2b8-092b-4091-8a21-f5166c29ae9b" containerName="glance-log" containerID="cri-o://1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472" gracePeriod=30 Nov 24 10:20:46 crc kubenswrapper[4944]: I1124 10:20:46.644097 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.644083472 podStartE2EDuration="2.644083472s" podCreationTimestamp="2025-11-24 10:20:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:20:46.64341896 +0000 UTC m=+5307.177859432" watchObservedRunningTime="2025-11-24 10:20:46.644083472 +0000 UTC m=+5307.178523934" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.184361 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.371570 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-combined-ca-bundle\") pod \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.371655 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-logs\") pod \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.371806 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-ceph\") pod \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.371849 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-config-data\") pod \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.371866 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-scripts\") pod \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.371901 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-httpd-run\") pod \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.371924 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs2lv\" (UniqueName: \"kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-kube-api-access-hs2lv\") pod \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\" (UID: \"48e5d2b8-092b-4091-8a21-f5166c29ae9b\") " Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.372176 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-logs" (OuterVolumeSpecName: "logs") pod "48e5d2b8-092b-4091-8a21-f5166c29ae9b" (UID: "48e5d2b8-092b-4091-8a21-f5166c29ae9b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.372259 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "48e5d2b8-092b-4091-8a21-f5166c29ae9b" (UID: "48e5d2b8-092b-4091-8a21-f5166c29ae9b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.372325 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.376565 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-ceph" (OuterVolumeSpecName: "ceph") pod "48e5d2b8-092b-4091-8a21-f5166c29ae9b" (UID: "48e5d2b8-092b-4091-8a21-f5166c29ae9b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.376964 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-kube-api-access-hs2lv" (OuterVolumeSpecName: "kube-api-access-hs2lv") pod "48e5d2b8-092b-4091-8a21-f5166c29ae9b" (UID: "48e5d2b8-092b-4091-8a21-f5166c29ae9b"). InnerVolumeSpecName "kube-api-access-hs2lv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.377695 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-scripts" (OuterVolumeSpecName: "scripts") pod "48e5d2b8-092b-4091-8a21-f5166c29ae9b" (UID: "48e5d2b8-092b-4091-8a21-f5166c29ae9b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.400688 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48e5d2b8-092b-4091-8a21-f5166c29ae9b" (UID: "48e5d2b8-092b-4091-8a21-f5166c29ae9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.419511 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-config-data" (OuterVolumeSpecName: "config-data") pod "48e5d2b8-092b-4091-8a21-f5166c29ae9b" (UID: "48e5d2b8-092b-4091-8a21-f5166c29ae9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.474156 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.474199 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.474217 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/48e5d2b8-092b-4091-8a21-f5166c29ae9b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.474234 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs2lv\" (UniqueName: \"kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-kube-api-access-hs2lv\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.474831 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e5d2b8-092b-4091-8a21-f5166c29ae9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.475030 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/48e5d2b8-092b-4091-8a21-f5166c29ae9b-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.631308 4944 generic.go:334] "Generic (PLEG): container finished" podID="48e5d2b8-092b-4091-8a21-f5166c29ae9b" containerID="a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16" exitCode=0 Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.631344 4944 generic.go:334] "Generic (PLEG): container finished" podID="48e5d2b8-092b-4091-8a21-f5166c29ae9b" containerID="1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472" exitCode=143 Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.631363 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.631355 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48e5d2b8-092b-4091-8a21-f5166c29ae9b","Type":"ContainerDied","Data":"a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16"} Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.631628 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48e5d2b8-092b-4091-8a21-f5166c29ae9b","Type":"ContainerDied","Data":"1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472"} Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.631675 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"48e5d2b8-092b-4091-8a21-f5166c29ae9b","Type":"ContainerDied","Data":"9f9587ac15e41f04786dd5931458c522634e538885895536838cef8716beb2bc"} Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.631696 4944 scope.go:117] "RemoveContainer" containerID="a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.661001 4944 scope.go:117] "RemoveContainer" containerID="1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.667093 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.687069 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.692396 4944 scope.go:117] "RemoveContainer" containerID="a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16" Nov 24 10:20:47 crc kubenswrapper[4944]: E1124 10:20:47.692862 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16\": container with ID starting with a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16 not found: ID does not exist" containerID="a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.692893 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16"} err="failed to get container status \"a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16\": rpc error: code = NotFound desc = could not find container \"a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16\": container with ID starting with a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16 not found: ID does not exist" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.692914 4944 scope.go:117] "RemoveContainer" containerID="1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472" Nov 24 10:20:47 crc kubenswrapper[4944]: E1124 10:20:47.693384 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472\": container with ID starting with 1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472 not found: ID does not exist" containerID="1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.693432 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472"} err="failed to get container status \"1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472\": rpc error: code = NotFound desc = could not find container \"1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472\": container with ID starting with 1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472 not found: ID does not exist" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.693460 4944 scope.go:117] "RemoveContainer" containerID="a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.693828 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16"} err="failed to get container status \"a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16\": rpc error: code = NotFound desc = could not find container \"a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16\": container with ID starting with a94311a1c8ae989968bc39a7eccd3b7b6158f052cb630c9f3dcab385ee7a3b16 not found: ID does not exist" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.693852 4944 scope.go:117] "RemoveContainer" containerID="1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.694137 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472"} err="failed to get container status \"1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472\": rpc error: code = NotFound desc = could not find container \"1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472\": container with ID starting with 1a3748b850e0096aaf8b831e26c28034a61ac5a1a2cd51f56fccb56dae29b472 not found: ID does not exist" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.695069 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:20:47 crc kubenswrapper[4944]: E1124 10:20:47.695487 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48e5d2b8-092b-4091-8a21-f5166c29ae9b" containerName="glance-httpd" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.695506 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="48e5d2b8-092b-4091-8a21-f5166c29ae9b" containerName="glance-httpd" Nov 24 10:20:47 crc kubenswrapper[4944]: E1124 10:20:47.695536 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48e5d2b8-092b-4091-8a21-f5166c29ae9b" containerName="glance-log" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.695544 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="48e5d2b8-092b-4091-8a21-f5166c29ae9b" containerName="glance-log" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.695743 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="48e5d2b8-092b-4091-8a21-f5166c29ae9b" containerName="glance-log" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.695774 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="48e5d2b8-092b-4091-8a21-f5166c29ae9b" containerName="glance-httpd" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.696868 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.718742 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.719534 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.881950 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-logs\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.882196 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.882273 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.882347 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.882372 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5vmk\" (UniqueName: \"kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-kube-api-access-c5vmk\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.882542 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.882569 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.983647 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-logs\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.983739 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.983764 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.983785 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.983805 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5vmk\" (UniqueName: \"kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-kube-api-access-c5vmk\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.983847 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.983863 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.984215 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-logs\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.984268 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.988127 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-ceph\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.988290 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.988864 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.989553 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:47 crc kubenswrapper[4944]: I1124 10:20:47.999509 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5vmk\" (UniqueName: \"kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-kube-api-access-c5vmk\") pod \"glance-default-internal-api-0\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:20:48 crc kubenswrapper[4944]: I1124 10:20:48.057891 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:48 crc kubenswrapper[4944]: I1124 10:20:48.293832 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48e5d2b8-092b-4091-8a21-f5166c29ae9b" path="/var/lib/kubelet/pods/48e5d2b8-092b-4091-8a21-f5166c29ae9b/volumes" Nov 24 10:20:48 crc kubenswrapper[4944]: I1124 10:20:48.562247 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:20:48 crc kubenswrapper[4944]: I1124 10:20:48.641377 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f","Type":"ContainerStarted","Data":"eaaa45491b18af4a02799d3f201bc5a8d89ef6da12e0cf9f1ffff743305ab6ac"} Nov 24 10:20:49 crc kubenswrapper[4944]: I1124 10:20:49.652091 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f","Type":"ContainerStarted","Data":"2dd1a505008d8acd50a78166cca4c0e762847d1f8ebcf1e93189444b348e6485"} Nov 24 10:20:49 crc kubenswrapper[4944]: I1124 10:20:49.652441 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f","Type":"ContainerStarted","Data":"b87f06c33d5fb6ea6d1cdf4f6a4978d6bddfebcfd760111f2e11997ada543bbe"} Nov 24 10:20:49 crc kubenswrapper[4944]: I1124 10:20:49.675105 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.675085194 podStartE2EDuration="2.675085194s" podCreationTimestamp="2025-11-24 10:20:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:20:49.669727213 +0000 UTC m=+5310.204167685" watchObservedRunningTime="2025-11-24 10:20:49.675085194 +0000 UTC m=+5310.209525666" Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.256695 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.317028 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fcbf9c88f-tdpxb"] Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.317333 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" podUID="2b1c4205-8dbd-422a-b35f-6c1e4a463c21" containerName="dnsmasq-dns" containerID="cri-o://a3f25262dc115a792d485a0eb7d390d136ca5a27af9fe6ac160460ef0fea96a9" gracePeriod=10 Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.680025 4944 generic.go:334] "Generic (PLEG): container finished" podID="2b1c4205-8dbd-422a-b35f-6c1e4a463c21" containerID="a3f25262dc115a792d485a0eb7d390d136ca5a27af9fe6ac160460ef0fea96a9" exitCode=0 Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.680459 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" event={"ID":"2b1c4205-8dbd-422a-b35f-6c1e4a463c21","Type":"ContainerDied","Data":"a3f25262dc115a792d485a0eb7d390d136ca5a27af9fe6ac160460ef0fea96a9"} Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.787722 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.973737 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-sb\") pod \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.973819 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-dns-svc\") pod \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.973860 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-config\") pod \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.973886 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nb2hr\" (UniqueName: \"kubernetes.io/projected/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-kube-api-access-nb2hr\") pod \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.974023 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-nb\") pod \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\" (UID: \"2b1c4205-8dbd-422a-b35f-6c1e4a463c21\") " Nov 24 10:20:52 crc kubenswrapper[4944]: I1124 10:20:52.995237 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-kube-api-access-nb2hr" (OuterVolumeSpecName: "kube-api-access-nb2hr") pod "2b1c4205-8dbd-422a-b35f-6c1e4a463c21" (UID: "2b1c4205-8dbd-422a-b35f-6c1e4a463c21"). InnerVolumeSpecName "kube-api-access-nb2hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.021881 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2b1c4205-8dbd-422a-b35f-6c1e4a463c21" (UID: "2b1c4205-8dbd-422a-b35f-6c1e4a463c21"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.023161 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2b1c4205-8dbd-422a-b35f-6c1e4a463c21" (UID: "2b1c4205-8dbd-422a-b35f-6c1e4a463c21"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.029791 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2b1c4205-8dbd-422a-b35f-6c1e4a463c21" (UID: "2b1c4205-8dbd-422a-b35f-6c1e4a463c21"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.032656 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-config" (OuterVolumeSpecName: "config") pod "2b1c4205-8dbd-422a-b35f-6c1e4a463c21" (UID: "2b1c4205-8dbd-422a-b35f-6c1e4a463c21"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.077192 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.077231 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.077241 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.077251 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.077262 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nb2hr\" (UniqueName: \"kubernetes.io/projected/2b1c4205-8dbd-422a-b35f-6c1e4a463c21-kube-api-access-nb2hr\") on node \"crc\" DevicePath \"\"" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.688870 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" event={"ID":"2b1c4205-8dbd-422a-b35f-6c1e4a463c21","Type":"ContainerDied","Data":"6ac3b97fca202eec40e950899a48669af9b3ff1d650c9dc9246b32978e0f1b5b"} Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.688934 4944 scope.go:117] "RemoveContainer" containerID="a3f25262dc115a792d485a0eb7d390d136ca5a27af9fe6ac160460ef0fea96a9" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.688980 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fcbf9c88f-tdpxb" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.707331 4944 scope.go:117] "RemoveContainer" containerID="64dfa47e9a59afaac2acd686c2db629e4a5a14facc81c5b51507600349a02562" Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.730920 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fcbf9c88f-tdpxb"] Nov 24 10:20:53 crc kubenswrapper[4944]: I1124 10:20:53.736656 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fcbf9c88f-tdpxb"] Nov 24 10:20:54 crc kubenswrapper[4944]: I1124 10:20:54.285871 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b1c4205-8dbd-422a-b35f-6c1e4a463c21" path="/var/lib/kubelet/pods/2b1c4205-8dbd-422a-b35f-6c1e4a463c21/volumes" Nov 24 10:20:55 crc kubenswrapper[4944]: I1124 10:20:55.004695 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 10:20:55 crc kubenswrapper[4944]: I1124 10:20:55.004791 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 10:20:55 crc kubenswrapper[4944]: I1124 10:20:55.041573 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 10:20:55 crc kubenswrapper[4944]: I1124 10:20:55.043525 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 10:20:55 crc kubenswrapper[4944]: I1124 10:20:55.708262 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 10:20:55 crc kubenswrapper[4944]: I1124 10:20:55.708315 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 10:20:57 crc kubenswrapper[4944]: I1124 10:20:57.698519 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 10:20:57 crc kubenswrapper[4944]: I1124 10:20:57.718876 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 10:20:58 crc kubenswrapper[4944]: I1124 10:20:58.058987 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:58 crc kubenswrapper[4944]: I1124 10:20:58.059293 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:58 crc kubenswrapper[4944]: I1124 10:20:58.093931 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:58 crc kubenswrapper[4944]: I1124 10:20:58.098419 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:58 crc kubenswrapper[4944]: I1124 10:20:58.730860 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:58 crc kubenswrapper[4944]: I1124 10:20:58.730950 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.647487 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cl99t"] Nov 24 10:20:59 crc kubenswrapper[4944]: E1124 10:20:59.647942 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b1c4205-8dbd-422a-b35f-6c1e4a463c21" containerName="dnsmasq-dns" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.647966 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b1c4205-8dbd-422a-b35f-6c1e4a463c21" containerName="dnsmasq-dns" Nov 24 10:20:59 crc kubenswrapper[4944]: E1124 10:20:59.647980 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b1c4205-8dbd-422a-b35f-6c1e4a463c21" containerName="init" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.647988 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b1c4205-8dbd-422a-b35f-6c1e4a463c21" containerName="init" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.648220 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b1c4205-8dbd-422a-b35f-6c1e4a463c21" containerName="dnsmasq-dns" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.649828 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.657072 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cl99t"] Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.696994 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w8rn\" (UniqueName: \"kubernetes.io/projected/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-kube-api-access-8w8rn\") pod \"redhat-operators-cl99t\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.697129 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-catalog-content\") pod \"redhat-operators-cl99t\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.697152 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-utilities\") pod \"redhat-operators-cl99t\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.798312 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w8rn\" (UniqueName: \"kubernetes.io/projected/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-kube-api-access-8w8rn\") pod \"redhat-operators-cl99t\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.798476 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-catalog-content\") pod \"redhat-operators-cl99t\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.798495 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-utilities\") pod \"redhat-operators-cl99t\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.799917 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-catalog-content\") pod \"redhat-operators-cl99t\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.799966 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-utilities\") pod \"redhat-operators-cl99t\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.816137 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w8rn\" (UniqueName: \"kubernetes.io/projected/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-kube-api-access-8w8rn\") pod \"redhat-operators-cl99t\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:20:59 crc kubenswrapper[4944]: I1124 10:20:59.971881 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:21:00 crc kubenswrapper[4944]: I1124 10:21:00.430986 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cl99t"] Nov 24 10:21:00 crc kubenswrapper[4944]: W1124 10:21:00.436088 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd767d4ae_6eb4_4fff_a5eb_c0d08b5b0689.slice/crio-1a3a47146eee6e437cc768b6bd8cd3e18dfe1006fb3431a3b8b7479a02feab8f WatchSource:0}: Error finding container 1a3a47146eee6e437cc768b6bd8cd3e18dfe1006fb3431a3b8b7479a02feab8f: Status 404 returned error can't find the container with id 1a3a47146eee6e437cc768b6bd8cd3e18dfe1006fb3431a3b8b7479a02feab8f Nov 24 10:21:00 crc kubenswrapper[4944]: I1124 10:21:00.746626 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 10:21:00 crc kubenswrapper[4944]: I1124 10:21:00.746906 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl99t" event={"ID":"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689","Type":"ContainerDied","Data":"872efa655bef63c716fb3debc537137f154ec440864e0352d2bdacd147bdb1da"} Nov 24 10:21:00 crc kubenswrapper[4944]: I1124 10:21:00.746721 4944 generic.go:334] "Generic (PLEG): container finished" podID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" containerID="872efa655bef63c716fb3debc537137f154ec440864e0352d2bdacd147bdb1da" exitCode=0 Nov 24 10:21:00 crc kubenswrapper[4944]: I1124 10:21:00.747175 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 10:21:00 crc kubenswrapper[4944]: I1124 10:21:00.747173 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl99t" event={"ID":"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689","Type":"ContainerStarted","Data":"1a3a47146eee6e437cc768b6bd8cd3e18dfe1006fb3431a3b8b7479a02feab8f"} Nov 24 10:21:00 crc kubenswrapper[4944]: I1124 10:21:00.767069 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 10:21:01 crc kubenswrapper[4944]: I1124 10:21:01.756255 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl99t" event={"ID":"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689","Type":"ContainerStarted","Data":"3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0"} Nov 24 10:21:02 crc kubenswrapper[4944]: I1124 10:21:02.768123 4944 generic.go:334] "Generic (PLEG): container finished" podID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" containerID="3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0" exitCode=0 Nov 24 10:21:02 crc kubenswrapper[4944]: I1124 10:21:02.768169 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl99t" event={"ID":"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689","Type":"ContainerDied","Data":"3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0"} Nov 24 10:21:03 crc kubenswrapper[4944]: I1124 10:21:03.777442 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl99t" event={"ID":"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689","Type":"ContainerStarted","Data":"36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e"} Nov 24 10:21:03 crc kubenswrapper[4944]: I1124 10:21:03.798908 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cl99t" podStartSLOduration=2.313844451 podStartE2EDuration="4.79888892s" podCreationTimestamp="2025-11-24 10:20:59 +0000 UTC" firstStartedPulling="2025-11-24 10:21:00.748134809 +0000 UTC m=+5321.282575271" lastFinishedPulling="2025-11-24 10:21:03.233179288 +0000 UTC m=+5323.767619740" observedRunningTime="2025-11-24 10:21:03.79385608 +0000 UTC m=+5324.328296552" watchObservedRunningTime="2025-11-24 10:21:03.79888892 +0000 UTC m=+5324.333329382" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.405869 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-27fnl"] Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.407305 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-27fnl" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.415537 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-27fnl"] Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.508322 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-74df-account-create-dmhxf"] Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.509534 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwc6v\" (UniqueName: \"kubernetes.io/projected/23161bdc-eb83-4926-a369-8a61d912b308-kube-api-access-kwc6v\") pod \"placement-db-create-27fnl\" (UID: \"23161bdc-eb83-4926-a369-8a61d912b308\") " pod="openstack/placement-db-create-27fnl" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.509597 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23161bdc-eb83-4926-a369-8a61d912b308-operator-scripts\") pod \"placement-db-create-27fnl\" (UID: \"23161bdc-eb83-4926-a369-8a61d912b308\") " pod="openstack/placement-db-create-27fnl" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.510533 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-74df-account-create-dmhxf" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.513206 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.522814 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-74df-account-create-dmhxf"] Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.616412 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97s7v\" (UniqueName: \"kubernetes.io/projected/ab617caa-207c-48b4-a541-32b97f4a7a53-kube-api-access-97s7v\") pod \"placement-74df-account-create-dmhxf\" (UID: \"ab617caa-207c-48b4-a541-32b97f4a7a53\") " pod="openstack/placement-74df-account-create-dmhxf" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.616470 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab617caa-207c-48b4-a541-32b97f4a7a53-operator-scripts\") pod \"placement-74df-account-create-dmhxf\" (UID: \"ab617caa-207c-48b4-a541-32b97f4a7a53\") " pod="openstack/placement-74df-account-create-dmhxf" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.616498 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwc6v\" (UniqueName: \"kubernetes.io/projected/23161bdc-eb83-4926-a369-8a61d912b308-kube-api-access-kwc6v\") pod \"placement-db-create-27fnl\" (UID: \"23161bdc-eb83-4926-a369-8a61d912b308\") " pod="openstack/placement-db-create-27fnl" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.616521 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23161bdc-eb83-4926-a369-8a61d912b308-operator-scripts\") pod \"placement-db-create-27fnl\" (UID: \"23161bdc-eb83-4926-a369-8a61d912b308\") " pod="openstack/placement-db-create-27fnl" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.617300 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23161bdc-eb83-4926-a369-8a61d912b308-operator-scripts\") pod \"placement-db-create-27fnl\" (UID: \"23161bdc-eb83-4926-a369-8a61d912b308\") " pod="openstack/placement-db-create-27fnl" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.635254 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwc6v\" (UniqueName: \"kubernetes.io/projected/23161bdc-eb83-4926-a369-8a61d912b308-kube-api-access-kwc6v\") pod \"placement-db-create-27fnl\" (UID: \"23161bdc-eb83-4926-a369-8a61d912b308\") " pod="openstack/placement-db-create-27fnl" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.717947 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97s7v\" (UniqueName: \"kubernetes.io/projected/ab617caa-207c-48b4-a541-32b97f4a7a53-kube-api-access-97s7v\") pod \"placement-74df-account-create-dmhxf\" (UID: \"ab617caa-207c-48b4-a541-32b97f4a7a53\") " pod="openstack/placement-74df-account-create-dmhxf" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.718001 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab617caa-207c-48b4-a541-32b97f4a7a53-operator-scripts\") pod \"placement-74df-account-create-dmhxf\" (UID: \"ab617caa-207c-48b4-a541-32b97f4a7a53\") " pod="openstack/placement-74df-account-create-dmhxf" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.718878 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab617caa-207c-48b4-a541-32b97f4a7a53-operator-scripts\") pod \"placement-74df-account-create-dmhxf\" (UID: \"ab617caa-207c-48b4-a541-32b97f4a7a53\") " pod="openstack/placement-74df-account-create-dmhxf" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.734961 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-27fnl" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.736198 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97s7v\" (UniqueName: \"kubernetes.io/projected/ab617caa-207c-48b4-a541-32b97f4a7a53-kube-api-access-97s7v\") pod \"placement-74df-account-create-dmhxf\" (UID: \"ab617caa-207c-48b4-a541-32b97f4a7a53\") " pod="openstack/placement-74df-account-create-dmhxf" Nov 24 10:21:06 crc kubenswrapper[4944]: I1124 10:21:06.852913 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-74df-account-create-dmhxf" Nov 24 10:21:07 crc kubenswrapper[4944]: I1124 10:21:07.223214 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-27fnl"] Nov 24 10:21:07 crc kubenswrapper[4944]: W1124 10:21:07.226466 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23161bdc_eb83_4926_a369_8a61d912b308.slice/crio-cd1509f8b35c2bb3b0673da0a54225bcd769396a41fb41a8fdcb8cc1257147b0 WatchSource:0}: Error finding container cd1509f8b35c2bb3b0673da0a54225bcd769396a41fb41a8fdcb8cc1257147b0: Status 404 returned error can't find the container with id cd1509f8b35c2bb3b0673da0a54225bcd769396a41fb41a8fdcb8cc1257147b0 Nov 24 10:21:07 crc kubenswrapper[4944]: I1124 10:21:07.326733 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-74df-account-create-dmhxf"] Nov 24 10:21:07 crc kubenswrapper[4944]: W1124 10:21:07.328859 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab617caa_207c_48b4_a541_32b97f4a7a53.slice/crio-87bca4ca3779388437f3dd976a3a6fac92e630f8ef957cdb2d44f1cf59538054 WatchSource:0}: Error finding container 87bca4ca3779388437f3dd976a3a6fac92e630f8ef957cdb2d44f1cf59538054: Status 404 returned error can't find the container with id 87bca4ca3779388437f3dd976a3a6fac92e630f8ef957cdb2d44f1cf59538054 Nov 24 10:21:07 crc kubenswrapper[4944]: I1124 10:21:07.808949 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74df-account-create-dmhxf" event={"ID":"ab617caa-207c-48b4-a541-32b97f4a7a53","Type":"ContainerStarted","Data":"7d243a0e8a59fd759d977f83018e79dd7a87fc7252e628966e29c4ea33148fff"} Nov 24 10:21:07 crc kubenswrapper[4944]: I1124 10:21:07.809000 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74df-account-create-dmhxf" event={"ID":"ab617caa-207c-48b4-a541-32b97f4a7a53","Type":"ContainerStarted","Data":"87bca4ca3779388437f3dd976a3a6fac92e630f8ef957cdb2d44f1cf59538054"} Nov 24 10:21:07 crc kubenswrapper[4944]: I1124 10:21:07.812380 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-27fnl" event={"ID":"23161bdc-eb83-4926-a369-8a61d912b308","Type":"ContainerStarted","Data":"d3dfac5e9eeb53b094a68347590f5d09a843ef8f8cf0d8cfeae53e2b7a56d184"} Nov 24 10:21:07 crc kubenswrapper[4944]: I1124 10:21:07.812460 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-27fnl" event={"ID":"23161bdc-eb83-4926-a369-8a61d912b308","Type":"ContainerStarted","Data":"cd1509f8b35c2bb3b0673da0a54225bcd769396a41fb41a8fdcb8cc1257147b0"} Nov 24 10:21:07 crc kubenswrapper[4944]: I1124 10:21:07.825781 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-74df-account-create-dmhxf" podStartSLOduration=1.825756693 podStartE2EDuration="1.825756693s" podCreationTimestamp="2025-11-24 10:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:21:07.823419769 +0000 UTC m=+5328.357860231" watchObservedRunningTime="2025-11-24 10:21:07.825756693 +0000 UTC m=+5328.360197155" Nov 24 10:21:07 crc kubenswrapper[4944]: I1124 10:21:07.839753 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-27fnl" podStartSLOduration=1.839735098 podStartE2EDuration="1.839735098s" podCreationTimestamp="2025-11-24 10:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:21:07.83726094 +0000 UTC m=+5328.371701422" watchObservedRunningTime="2025-11-24 10:21:07.839735098 +0000 UTC m=+5328.374175570" Nov 24 10:21:08 crc kubenswrapper[4944]: I1124 10:21:08.822924 4944 generic.go:334] "Generic (PLEG): container finished" podID="ab617caa-207c-48b4-a541-32b97f4a7a53" containerID="7d243a0e8a59fd759d977f83018e79dd7a87fc7252e628966e29c4ea33148fff" exitCode=0 Nov 24 10:21:08 crc kubenswrapper[4944]: I1124 10:21:08.823028 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74df-account-create-dmhxf" event={"ID":"ab617caa-207c-48b4-a541-32b97f4a7a53","Type":"ContainerDied","Data":"7d243a0e8a59fd759d977f83018e79dd7a87fc7252e628966e29c4ea33148fff"} Nov 24 10:21:08 crc kubenswrapper[4944]: I1124 10:21:08.825247 4944 generic.go:334] "Generic (PLEG): container finished" podID="23161bdc-eb83-4926-a369-8a61d912b308" containerID="d3dfac5e9eeb53b094a68347590f5d09a843ef8f8cf0d8cfeae53e2b7a56d184" exitCode=0 Nov 24 10:21:08 crc kubenswrapper[4944]: I1124 10:21:08.825297 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-27fnl" event={"ID":"23161bdc-eb83-4926-a369-8a61d912b308","Type":"ContainerDied","Data":"d3dfac5e9eeb53b094a68347590f5d09a843ef8f8cf0d8cfeae53e2b7a56d184"} Nov 24 10:21:09 crc kubenswrapper[4944]: I1124 10:21:09.972105 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:21:09 crc kubenswrapper[4944]: I1124 10:21:09.972151 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.021760 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.226216 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-27fnl" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.233453 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-74df-account-create-dmhxf" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.379755 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97s7v\" (UniqueName: \"kubernetes.io/projected/ab617caa-207c-48b4-a541-32b97f4a7a53-kube-api-access-97s7v\") pod \"ab617caa-207c-48b4-a541-32b97f4a7a53\" (UID: \"ab617caa-207c-48b4-a541-32b97f4a7a53\") " Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.379832 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwc6v\" (UniqueName: \"kubernetes.io/projected/23161bdc-eb83-4926-a369-8a61d912b308-kube-api-access-kwc6v\") pod \"23161bdc-eb83-4926-a369-8a61d912b308\" (UID: \"23161bdc-eb83-4926-a369-8a61d912b308\") " Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.379869 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab617caa-207c-48b4-a541-32b97f4a7a53-operator-scripts\") pod \"ab617caa-207c-48b4-a541-32b97f4a7a53\" (UID: \"ab617caa-207c-48b4-a541-32b97f4a7a53\") " Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.379911 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23161bdc-eb83-4926-a369-8a61d912b308-operator-scripts\") pod \"23161bdc-eb83-4926-a369-8a61d912b308\" (UID: \"23161bdc-eb83-4926-a369-8a61d912b308\") " Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.380541 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab617caa-207c-48b4-a541-32b97f4a7a53-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ab617caa-207c-48b4-a541-32b97f4a7a53" (UID: "ab617caa-207c-48b4-a541-32b97f4a7a53"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.380656 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23161bdc-eb83-4926-a369-8a61d912b308-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "23161bdc-eb83-4926-a369-8a61d912b308" (UID: "23161bdc-eb83-4926-a369-8a61d912b308"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.385525 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23161bdc-eb83-4926-a369-8a61d912b308-kube-api-access-kwc6v" (OuterVolumeSpecName: "kube-api-access-kwc6v") pod "23161bdc-eb83-4926-a369-8a61d912b308" (UID: "23161bdc-eb83-4926-a369-8a61d912b308"). InnerVolumeSpecName "kube-api-access-kwc6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.386341 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab617caa-207c-48b4-a541-32b97f4a7a53-kube-api-access-97s7v" (OuterVolumeSpecName: "kube-api-access-97s7v") pod "ab617caa-207c-48b4-a541-32b97f4a7a53" (UID: "ab617caa-207c-48b4-a541-32b97f4a7a53"). InnerVolumeSpecName "kube-api-access-97s7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.482169 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97s7v\" (UniqueName: \"kubernetes.io/projected/ab617caa-207c-48b4-a541-32b97f4a7a53-kube-api-access-97s7v\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.482213 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwc6v\" (UniqueName: \"kubernetes.io/projected/23161bdc-eb83-4926-a369-8a61d912b308-kube-api-access-kwc6v\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.482227 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab617caa-207c-48b4-a541-32b97f4a7a53-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.482239 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23161bdc-eb83-4926-a369-8a61d912b308-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.843559 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74df-account-create-dmhxf" event={"ID":"ab617caa-207c-48b4-a541-32b97f4a7a53","Type":"ContainerDied","Data":"87bca4ca3779388437f3dd976a3a6fac92e630f8ef957cdb2d44f1cf59538054"} Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.843999 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87bca4ca3779388437f3dd976a3a6fac92e630f8ef957cdb2d44f1cf59538054" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.843583 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-74df-account-create-dmhxf" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.846668 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-27fnl" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.846663 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-27fnl" event={"ID":"23161bdc-eb83-4926-a369-8a61d912b308","Type":"ContainerDied","Data":"cd1509f8b35c2bb3b0673da0a54225bcd769396a41fb41a8fdcb8cc1257147b0"} Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.846739 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd1509f8b35c2bb3b0673da0a54225bcd769396a41fb41a8fdcb8cc1257147b0" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.898803 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:21:10 crc kubenswrapper[4944]: I1124 10:21:10.950681 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cl99t"] Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.753319 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f7979997-bz56v"] Nov 24 10:21:11 crc kubenswrapper[4944]: E1124 10:21:11.753804 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab617caa-207c-48b4-a541-32b97f4a7a53" containerName="mariadb-account-create" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.753822 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab617caa-207c-48b4-a541-32b97f4a7a53" containerName="mariadb-account-create" Nov 24 10:21:11 crc kubenswrapper[4944]: E1124 10:21:11.753836 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23161bdc-eb83-4926-a369-8a61d912b308" containerName="mariadb-database-create" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.753844 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="23161bdc-eb83-4926-a369-8a61d912b308" containerName="mariadb-database-create" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.754096 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab617caa-207c-48b4-a541-32b97f4a7a53" containerName="mariadb-account-create" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.754127 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="23161bdc-eb83-4926-a369-8a61d912b308" containerName="mariadb-database-create" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.755331 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.761967 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f7979997-bz56v"] Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.796345 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-p2p5g"] Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.797772 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.804200 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.804564 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.804628 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-x5j4d" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.824071 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-p2p5g"] Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.903510 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-nb\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.903766 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-config\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.903806 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-sb\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.903920 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqw78\" (UniqueName: \"kubernetes.io/projected/5e324380-8997-48e9-acb6-b7c22e9eecba-kube-api-access-gqw78\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.903958 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-combined-ca-bundle\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.904043 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-scripts\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.904138 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-config-data\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.904177 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzp9c\" (UniqueName: \"kubernetes.io/projected/7b566b74-289c-4e74-9708-01730c66c939-kube-api-access-rzp9c\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.904202 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-dns-svc\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:11 crc kubenswrapper[4944]: I1124 10:21:11.904222 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e324380-8997-48e9-acb6-b7c22e9eecba-logs\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.005443 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqw78\" (UniqueName: \"kubernetes.io/projected/5e324380-8997-48e9-acb6-b7c22e9eecba-kube-api-access-gqw78\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.005497 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-combined-ca-bundle\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.005551 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-scripts\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.005602 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-config-data\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.005631 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzp9c\" (UniqueName: \"kubernetes.io/projected/7b566b74-289c-4e74-9708-01730c66c939-kube-api-access-rzp9c\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.005653 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-dns-svc\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.005680 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e324380-8997-48e9-acb6-b7c22e9eecba-logs\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.005725 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-nb\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.005776 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-config\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.005827 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-sb\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.006785 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e324380-8997-48e9-acb6-b7c22e9eecba-logs\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.007720 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-sb\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.008002 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-config\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.009550 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-nb\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.010980 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-config-data\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.012319 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-scripts\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.012395 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-dns-svc\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.014613 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-combined-ca-bundle\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.020742 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqw78\" (UniqueName: \"kubernetes.io/projected/5e324380-8997-48e9-acb6-b7c22e9eecba-kube-api-access-gqw78\") pod \"placement-db-sync-p2p5g\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.030553 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzp9c\" (UniqueName: \"kubernetes.io/projected/7b566b74-289c-4e74-9708-01730c66c939-kube-api-access-rzp9c\") pod \"dnsmasq-dns-f7979997-bz56v\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.074098 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.141929 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.545088 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f7979997-bz56v"] Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.620641 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-p2p5g"] Nov 24 10:21:12 crc kubenswrapper[4944]: W1124 10:21:12.624107 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e324380_8997_48e9_acb6_b7c22e9eecba.slice/crio-cb8300ceec88b63b6187c75be1f742aea9a87cbe57b1bf26736724f8f390b46a WatchSource:0}: Error finding container cb8300ceec88b63b6187c75be1f742aea9a87cbe57b1bf26736724f8f390b46a: Status 404 returned error can't find the container with id cb8300ceec88b63b6187c75be1f742aea9a87cbe57b1bf26736724f8f390b46a Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.865508 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7979997-bz56v" event={"ID":"7b566b74-289c-4e74-9708-01730c66c939","Type":"ContainerStarted","Data":"b47cfba1043ca125d51c9298923f72f3942012e0463ff03cf70138ea11bc5aea"} Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.866871 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-p2p5g" event={"ID":"5e324380-8997-48e9-acb6-b7c22e9eecba","Type":"ContainerStarted","Data":"cb8300ceec88b63b6187c75be1f742aea9a87cbe57b1bf26736724f8f390b46a"} Nov 24 10:21:12 crc kubenswrapper[4944]: I1124 10:21:12.866989 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cl99t" podUID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" containerName="registry-server" containerID="cri-o://36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e" gracePeriod=2 Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.309964 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.431941 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w8rn\" (UniqueName: \"kubernetes.io/projected/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-kube-api-access-8w8rn\") pod \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.432660 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-catalog-content\") pod \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.432785 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-utilities\") pod \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\" (UID: \"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689\") " Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.433538 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-utilities" (OuterVolumeSpecName: "utilities") pod "d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" (UID: "d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.436965 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-kube-api-access-8w8rn" (OuterVolumeSpecName: "kube-api-access-8w8rn") pod "d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" (UID: "d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689"). InnerVolumeSpecName "kube-api-access-8w8rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.515303 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" (UID: "d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.541000 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.541041 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.541068 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w8rn\" (UniqueName: \"kubernetes.io/projected/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689-kube-api-access-8w8rn\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.876961 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-p2p5g" event={"ID":"5e324380-8997-48e9-acb6-b7c22e9eecba","Type":"ContainerStarted","Data":"1e3fb39ec669308fd2146d8b1b299514c212cc268a17629dad96dd8bd9616433"} Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.880016 4944 generic.go:334] "Generic (PLEG): container finished" podID="7b566b74-289c-4e74-9708-01730c66c939" containerID="4fe194bc8caa34e0fbff1faaea361298d8c70c845d53bf961c5512922c724953" exitCode=0 Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.880110 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7979997-bz56v" event={"ID":"7b566b74-289c-4e74-9708-01730c66c939","Type":"ContainerDied","Data":"4fe194bc8caa34e0fbff1faaea361298d8c70c845d53bf961c5512922c724953"} Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.882787 4944 generic.go:334] "Generic (PLEG): container finished" podID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" containerID="36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e" exitCode=0 Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.882834 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl99t" event={"ID":"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689","Type":"ContainerDied","Data":"36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e"} Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.882864 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cl99t" event={"ID":"d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689","Type":"ContainerDied","Data":"1a3a47146eee6e437cc768b6bd8cd3e18dfe1006fb3431a3b8b7479a02feab8f"} Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.882884 4944 scope.go:117] "RemoveContainer" containerID="36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e" Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.883058 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cl99t" Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.894766 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-p2p5g" podStartSLOduration=2.89474228 podStartE2EDuration="2.89474228s" podCreationTimestamp="2025-11-24 10:21:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:21:13.890424253 +0000 UTC m=+5334.424864735" watchObservedRunningTime="2025-11-24 10:21:13.89474228 +0000 UTC m=+5334.429182752" Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.944111 4944 scope.go:117] "RemoveContainer" containerID="3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0" Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.952458 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cl99t"] Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.959966 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cl99t"] Nov 24 10:21:13 crc kubenswrapper[4944]: I1124 10:21:13.977663 4944 scope.go:117] "RemoveContainer" containerID="872efa655bef63c716fb3debc537137f154ec440864e0352d2bdacd147bdb1da" Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.001875 4944 scope.go:117] "RemoveContainer" containerID="36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e" Nov 24 10:21:14 crc kubenswrapper[4944]: E1124 10:21:14.002295 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e\": container with ID starting with 36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e not found: ID does not exist" containerID="36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e" Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.002405 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e"} err="failed to get container status \"36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e\": rpc error: code = NotFound desc = could not find container \"36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e\": container with ID starting with 36df483363af0da1b25cb00e247fcd66fd99ac297de9c4152988fdc0ada1b43e not found: ID does not exist" Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.002484 4944 scope.go:117] "RemoveContainer" containerID="3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0" Nov 24 10:21:14 crc kubenswrapper[4944]: E1124 10:21:14.002933 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0\": container with ID starting with 3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0 not found: ID does not exist" containerID="3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0" Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.002958 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0"} err="failed to get container status \"3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0\": rpc error: code = NotFound desc = could not find container \"3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0\": container with ID starting with 3cf0ef054ebe60fb722e42acca58b57c330852000305fa9f3b046d8249280da0 not found: ID does not exist" Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.002975 4944 scope.go:117] "RemoveContainer" containerID="872efa655bef63c716fb3debc537137f154ec440864e0352d2bdacd147bdb1da" Nov 24 10:21:14 crc kubenswrapper[4944]: E1124 10:21:14.003313 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"872efa655bef63c716fb3debc537137f154ec440864e0352d2bdacd147bdb1da\": container with ID starting with 872efa655bef63c716fb3debc537137f154ec440864e0352d2bdacd147bdb1da not found: ID does not exist" containerID="872efa655bef63c716fb3debc537137f154ec440864e0352d2bdacd147bdb1da" Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.003360 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"872efa655bef63c716fb3debc537137f154ec440864e0352d2bdacd147bdb1da"} err="failed to get container status \"872efa655bef63c716fb3debc537137f154ec440864e0352d2bdacd147bdb1da\": rpc error: code = NotFound desc = could not find container \"872efa655bef63c716fb3debc537137f154ec440864e0352d2bdacd147bdb1da\": container with ID starting with 872efa655bef63c716fb3debc537137f154ec440864e0352d2bdacd147bdb1da not found: ID does not exist" Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.288524 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" path="/var/lib/kubelet/pods/d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689/volumes" Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.902813 4944 generic.go:334] "Generic (PLEG): container finished" podID="5e324380-8997-48e9-acb6-b7c22e9eecba" containerID="1e3fb39ec669308fd2146d8b1b299514c212cc268a17629dad96dd8bd9616433" exitCode=0 Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.902906 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-p2p5g" event={"ID":"5e324380-8997-48e9-acb6-b7c22e9eecba","Type":"ContainerDied","Data":"1e3fb39ec669308fd2146d8b1b299514c212cc268a17629dad96dd8bd9616433"} Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.908186 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7979997-bz56v" event={"ID":"7b566b74-289c-4e74-9708-01730c66c939","Type":"ContainerStarted","Data":"70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf"} Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.908570 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:14 crc kubenswrapper[4944]: I1124 10:21:14.946229 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f7979997-bz56v" podStartSLOduration=3.946197081 podStartE2EDuration="3.946197081s" podCreationTimestamp="2025-11-24 10:21:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:21:14.937234935 +0000 UTC m=+5335.471675457" watchObservedRunningTime="2025-11-24 10:21:14.946197081 +0000 UTC m=+5335.480637583" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.086306 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9vj97"] Nov 24 10:21:16 crc kubenswrapper[4944]: E1124 10:21:16.088092 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" containerName="extract-content" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.088188 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" containerName="extract-content" Nov 24 10:21:16 crc kubenswrapper[4944]: E1124 10:21:16.088289 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" containerName="registry-server" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.088343 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" containerName="registry-server" Nov 24 10:21:16 crc kubenswrapper[4944]: E1124 10:21:16.088397 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" containerName="extract-utilities" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.088497 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" containerName="extract-utilities" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.088736 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d767d4ae-6eb4-4fff-a5eb-c0d08b5b0689" containerName="registry-server" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.092215 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.094630 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9vj97"] Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.194413 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgk4h\" (UniqueName: \"kubernetes.io/projected/80816194-484a-48d2-b749-b527341c4437-kube-api-access-qgk4h\") pod \"certified-operators-9vj97\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.194467 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-utilities\") pod \"certified-operators-9vj97\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.194604 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-catalog-content\") pod \"certified-operators-9vj97\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.267512 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.320328 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-catalog-content\") pod \"certified-operators-9vj97\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.320426 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgk4h\" (UniqueName: \"kubernetes.io/projected/80816194-484a-48d2-b749-b527341c4437-kube-api-access-qgk4h\") pod \"certified-operators-9vj97\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.320470 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-utilities\") pod \"certified-operators-9vj97\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.320846 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-catalog-content\") pod \"certified-operators-9vj97\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.320912 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-utilities\") pod \"certified-operators-9vj97\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.343948 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgk4h\" (UniqueName: \"kubernetes.io/projected/80816194-484a-48d2-b749-b527341c4437-kube-api-access-qgk4h\") pod \"certified-operators-9vj97\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.421888 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-config-data\") pod \"5e324380-8997-48e9-acb6-b7c22e9eecba\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.421932 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e324380-8997-48e9-acb6-b7c22e9eecba-logs\") pod \"5e324380-8997-48e9-acb6-b7c22e9eecba\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.421984 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-combined-ca-bundle\") pod \"5e324380-8997-48e9-acb6-b7c22e9eecba\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.422034 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqw78\" (UniqueName: \"kubernetes.io/projected/5e324380-8997-48e9-acb6-b7c22e9eecba-kube-api-access-gqw78\") pod \"5e324380-8997-48e9-acb6-b7c22e9eecba\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.422602 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-scripts\") pod \"5e324380-8997-48e9-acb6-b7c22e9eecba\" (UID: \"5e324380-8997-48e9-acb6-b7c22e9eecba\") " Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.422372 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e324380-8997-48e9-acb6-b7c22e9eecba-logs" (OuterVolumeSpecName: "logs") pod "5e324380-8997-48e9-acb6-b7c22e9eecba" (UID: "5e324380-8997-48e9-acb6-b7c22e9eecba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.425301 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e324380-8997-48e9-acb6-b7c22e9eecba-kube-api-access-gqw78" (OuterVolumeSpecName: "kube-api-access-gqw78") pod "5e324380-8997-48e9-acb6-b7c22e9eecba" (UID: "5e324380-8997-48e9-acb6-b7c22e9eecba"). InnerVolumeSpecName "kube-api-access-gqw78". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.433608 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-scripts" (OuterVolumeSpecName: "scripts") pod "5e324380-8997-48e9-acb6-b7c22e9eecba" (UID: "5e324380-8997-48e9-acb6-b7c22e9eecba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.433957 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.452971 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e324380-8997-48e9-acb6-b7c22e9eecba" (UID: "5e324380-8997-48e9-acb6-b7c22e9eecba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.458066 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-config-data" (OuterVolumeSpecName: "config-data") pod "5e324380-8997-48e9-acb6-b7c22e9eecba" (UID: "5e324380-8997-48e9-acb6-b7c22e9eecba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.524769 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.524805 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.524815 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e324380-8997-48e9-acb6-b7c22e9eecba-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.524824 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e324380-8997-48e9-acb6-b7c22e9eecba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.524834 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqw78\" (UniqueName: \"kubernetes.io/projected/5e324380-8997-48e9-acb6-b7c22e9eecba-kube-api-access-gqw78\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.707207 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9vj97"] Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.928898 4944 generic.go:334] "Generic (PLEG): container finished" podID="80816194-484a-48d2-b749-b527341c4437" containerID="98f9cdfb08de6689174f6ce74169a359a31d5c8d124604d0d324321676be8df4" exitCode=0 Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.929020 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vj97" event={"ID":"80816194-484a-48d2-b749-b527341c4437","Type":"ContainerDied","Data":"98f9cdfb08de6689174f6ce74169a359a31d5c8d124604d0d324321676be8df4"} Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.929122 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vj97" event={"ID":"80816194-484a-48d2-b749-b527341c4437","Type":"ContainerStarted","Data":"c01b86d1d688f95b67e7d725240272bfe6605284af59acb745ef31604a2a9f4f"} Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.931437 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-p2p5g" event={"ID":"5e324380-8997-48e9-acb6-b7c22e9eecba","Type":"ContainerDied","Data":"cb8300ceec88b63b6187c75be1f742aea9a87cbe57b1bf26736724f8f390b46a"} Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.931521 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb8300ceec88b63b6187c75be1f742aea9a87cbe57b1bf26736724f8f390b46a" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.931471 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-p2p5g" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.995498 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-774d97dd4-ks9d9"] Nov 24 10:21:16 crc kubenswrapper[4944]: E1124 10:21:16.995995 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e324380-8997-48e9-acb6-b7c22e9eecba" containerName="placement-db-sync" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.996013 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e324380-8997-48e9-acb6-b7c22e9eecba" containerName="placement-db-sync" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.996222 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e324380-8997-48e9-acb6-b7c22e9eecba" containerName="placement-db-sync" Nov 24 10:21:16 crc kubenswrapper[4944]: I1124 10:21:16.997415 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.002471 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-774d97dd4-ks9d9"] Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.008581 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.008674 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-x5j4d" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.008835 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.074855 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05fc0069-b41d-4157-8910-484784282a88-logs\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.074904 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05fc0069-b41d-4157-8910-484784282a88-config-data\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.074926 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05fc0069-b41d-4157-8910-484784282a88-combined-ca-bundle\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.074966 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05fc0069-b41d-4157-8910-484784282a88-scripts\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.075294 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv6kz\" (UniqueName: \"kubernetes.io/projected/05fc0069-b41d-4157-8910-484784282a88-kube-api-access-qv6kz\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.177431 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05fc0069-b41d-4157-8910-484784282a88-logs\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.177515 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05fc0069-b41d-4157-8910-484784282a88-config-data\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.177542 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05fc0069-b41d-4157-8910-484784282a88-combined-ca-bundle\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.177583 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05fc0069-b41d-4157-8910-484784282a88-scripts\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.177650 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv6kz\" (UniqueName: \"kubernetes.io/projected/05fc0069-b41d-4157-8910-484784282a88-kube-api-access-qv6kz\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.177927 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05fc0069-b41d-4157-8910-484784282a88-logs\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.183305 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05fc0069-b41d-4157-8910-484784282a88-scripts\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.183472 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05fc0069-b41d-4157-8910-484784282a88-config-data\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.187507 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05fc0069-b41d-4157-8910-484784282a88-combined-ca-bundle\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.193732 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv6kz\" (UniqueName: \"kubernetes.io/projected/05fc0069-b41d-4157-8910-484784282a88-kube-api-access-qv6kz\") pod \"placement-774d97dd4-ks9d9\" (UID: \"05fc0069-b41d-4157-8910-484784282a88\") " pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.318161 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.879796 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-774d97dd4-ks9d9"] Nov 24 10:21:17 crc kubenswrapper[4944]: W1124 10:21:17.886116 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05fc0069_b41d_4157_8910_484784282a88.slice/crio-0aabf1465be3665734842355737d65c06d1939289c233f6295d8f9e1b7ca61dd WatchSource:0}: Error finding container 0aabf1465be3665734842355737d65c06d1939289c233f6295d8f9e1b7ca61dd: Status 404 returned error can't find the container with id 0aabf1465be3665734842355737d65c06d1939289c233f6295d8f9e1b7ca61dd Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.940802 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-774d97dd4-ks9d9" event={"ID":"05fc0069-b41d-4157-8910-484784282a88","Type":"ContainerStarted","Data":"0aabf1465be3665734842355737d65c06d1939289c233f6295d8f9e1b7ca61dd"} Nov 24 10:21:17 crc kubenswrapper[4944]: I1124 10:21:17.942570 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vj97" event={"ID":"80816194-484a-48d2-b749-b527341c4437","Type":"ContainerStarted","Data":"e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a"} Nov 24 10:21:18 crc kubenswrapper[4944]: I1124 10:21:18.951194 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-774d97dd4-ks9d9" event={"ID":"05fc0069-b41d-4157-8910-484784282a88","Type":"ContainerStarted","Data":"62bc73e84650012bec1dedf9341885780a4cef345504cf214f3e1fdd2d579bd9"} Nov 24 10:21:18 crc kubenswrapper[4944]: I1124 10:21:18.951552 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:18 crc kubenswrapper[4944]: I1124 10:21:18.951567 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-774d97dd4-ks9d9" event={"ID":"05fc0069-b41d-4157-8910-484784282a88","Type":"ContainerStarted","Data":"522e8069a2b73df27f92839c36f9e9f97af0308c634f0a2ab15d3351a94a9c26"} Nov 24 10:21:18 crc kubenswrapper[4944]: I1124 10:21:18.953759 4944 generic.go:334] "Generic (PLEG): container finished" podID="80816194-484a-48d2-b749-b527341c4437" containerID="e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a" exitCode=0 Nov 24 10:21:18 crc kubenswrapper[4944]: I1124 10:21:18.953786 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vj97" event={"ID":"80816194-484a-48d2-b749-b527341c4437","Type":"ContainerDied","Data":"e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a"} Nov 24 10:21:18 crc kubenswrapper[4944]: I1124 10:21:18.990815 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-774d97dd4-ks9d9" podStartSLOduration=2.990796177 podStartE2EDuration="2.990796177s" podCreationTimestamp="2025-11-24 10:21:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:21:18.968367233 +0000 UTC m=+5339.502807685" watchObservedRunningTime="2025-11-24 10:21:18.990796177 +0000 UTC m=+5339.525236639" Nov 24 10:21:19 crc kubenswrapper[4944]: I1124 10:21:19.967030 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vj97" event={"ID":"80816194-484a-48d2-b749-b527341c4437","Type":"ContainerStarted","Data":"c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3"} Nov 24 10:21:19 crc kubenswrapper[4944]: I1124 10:21:19.967451 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:19 crc kubenswrapper[4944]: I1124 10:21:19.987693 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9vj97" podStartSLOduration=1.588080562 podStartE2EDuration="3.98767631s" podCreationTimestamp="2025-11-24 10:21:16 +0000 UTC" firstStartedPulling="2025-11-24 10:21:16.930823744 +0000 UTC m=+5337.465264206" lastFinishedPulling="2025-11-24 10:21:19.330419502 +0000 UTC m=+5339.864859954" observedRunningTime="2025-11-24 10:21:19.982877827 +0000 UTC m=+5340.517318299" watchObservedRunningTime="2025-11-24 10:21:19.98767631 +0000 UTC m=+5340.522116772" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.075894 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.122905 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64c4c7d56f-pzxnn"] Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.123448 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" podUID="f6aaa08f-11fd-495a-88f4-e1684d0f229d" containerName="dnsmasq-dns" containerID="cri-o://decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309" gracePeriod=10 Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.255761 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" podUID="f6aaa08f-11fd-495a-88f4-e1684d0f229d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.39:5353: connect: connection refused" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.628768 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.668113 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-nb\") pod \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.668306 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-dns-svc\") pod \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.668349 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-config\") pod \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.668404 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfxsj\" (UniqueName: \"kubernetes.io/projected/f6aaa08f-11fd-495a-88f4-e1684d0f229d-kube-api-access-hfxsj\") pod \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.668425 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-sb\") pod \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\" (UID: \"f6aaa08f-11fd-495a-88f4-e1684d0f229d\") " Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.674296 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6aaa08f-11fd-495a-88f4-e1684d0f229d-kube-api-access-hfxsj" (OuterVolumeSpecName: "kube-api-access-hfxsj") pod "f6aaa08f-11fd-495a-88f4-e1684d0f229d" (UID: "f6aaa08f-11fd-495a-88f4-e1684d0f229d"). InnerVolumeSpecName "kube-api-access-hfxsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.713452 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f6aaa08f-11fd-495a-88f4-e1684d0f229d" (UID: "f6aaa08f-11fd-495a-88f4-e1684d0f229d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.715617 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-config" (OuterVolumeSpecName: "config") pod "f6aaa08f-11fd-495a-88f4-e1684d0f229d" (UID: "f6aaa08f-11fd-495a-88f4-e1684d0f229d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.715845 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f6aaa08f-11fd-495a-88f4-e1684d0f229d" (UID: "f6aaa08f-11fd-495a-88f4-e1684d0f229d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.716406 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f6aaa08f-11fd-495a-88f4-e1684d0f229d" (UID: "f6aaa08f-11fd-495a-88f4-e1684d0f229d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.770423 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.770448 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.770458 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfxsj\" (UniqueName: \"kubernetes.io/projected/f6aaa08f-11fd-495a-88f4-e1684d0f229d-kube-api-access-hfxsj\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.770470 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.770478 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6aaa08f-11fd-495a-88f4-e1684d0f229d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.998766 4944 generic.go:334] "Generic (PLEG): container finished" podID="f6aaa08f-11fd-495a-88f4-e1684d0f229d" containerID="decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309" exitCode=0 Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.999153 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" event={"ID":"f6aaa08f-11fd-495a-88f4-e1684d0f229d","Type":"ContainerDied","Data":"decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309"} Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.999186 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" event={"ID":"f6aaa08f-11fd-495a-88f4-e1684d0f229d","Type":"ContainerDied","Data":"0a215c6dafaa81931fc9ffd730566bc28058891ddc269dbe1bfd14489b207a0c"} Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.999208 4944 scope.go:117] "RemoveContainer" containerID="decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309" Nov 24 10:21:22 crc kubenswrapper[4944]: I1124 10:21:22.999413 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64c4c7d56f-pzxnn" Nov 24 10:21:23 crc kubenswrapper[4944]: I1124 10:21:23.032937 4944 scope.go:117] "RemoveContainer" containerID="bab45faba9c74c08563981c1a099dbef2267aa6c1c11511dc7878fd1de5763bd" Nov 24 10:21:23 crc kubenswrapper[4944]: I1124 10:21:23.034867 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64c4c7d56f-pzxnn"] Nov 24 10:21:23 crc kubenswrapper[4944]: I1124 10:21:23.042446 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64c4c7d56f-pzxnn"] Nov 24 10:21:23 crc kubenswrapper[4944]: I1124 10:21:23.050736 4944 scope.go:117] "RemoveContainer" containerID="decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309" Nov 24 10:21:23 crc kubenswrapper[4944]: E1124 10:21:23.051350 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309\": container with ID starting with decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309 not found: ID does not exist" containerID="decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309" Nov 24 10:21:23 crc kubenswrapper[4944]: I1124 10:21:23.051386 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309"} err="failed to get container status \"decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309\": rpc error: code = NotFound desc = could not find container \"decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309\": container with ID starting with decb0821897976de8f4921031fff2b82a6583498f6acd3785fe953629d210309 not found: ID does not exist" Nov 24 10:21:23 crc kubenswrapper[4944]: I1124 10:21:23.051410 4944 scope.go:117] "RemoveContainer" containerID="bab45faba9c74c08563981c1a099dbef2267aa6c1c11511dc7878fd1de5763bd" Nov 24 10:21:23 crc kubenswrapper[4944]: E1124 10:21:23.051720 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bab45faba9c74c08563981c1a099dbef2267aa6c1c11511dc7878fd1de5763bd\": container with ID starting with bab45faba9c74c08563981c1a099dbef2267aa6c1c11511dc7878fd1de5763bd not found: ID does not exist" containerID="bab45faba9c74c08563981c1a099dbef2267aa6c1c11511dc7878fd1de5763bd" Nov 24 10:21:23 crc kubenswrapper[4944]: I1124 10:21:23.051751 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bab45faba9c74c08563981c1a099dbef2267aa6c1c11511dc7878fd1de5763bd"} err="failed to get container status \"bab45faba9c74c08563981c1a099dbef2267aa6c1c11511dc7878fd1de5763bd\": rpc error: code = NotFound desc = could not find container \"bab45faba9c74c08563981c1a099dbef2267aa6c1c11511dc7878fd1de5763bd\": container with ID starting with bab45faba9c74c08563981c1a099dbef2267aa6c1c11511dc7878fd1de5763bd not found: ID does not exist" Nov 24 10:21:24 crc kubenswrapper[4944]: I1124 10:21:24.286991 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6aaa08f-11fd-495a-88f4-e1684d0f229d" path="/var/lib/kubelet/pods/f6aaa08f-11fd-495a-88f4-e1684d0f229d/volumes" Nov 24 10:21:26 crc kubenswrapper[4944]: I1124 10:21:26.434890 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:26 crc kubenswrapper[4944]: I1124 10:21:26.435242 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:26 crc kubenswrapper[4944]: I1124 10:21:26.481959 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:27 crc kubenswrapper[4944]: I1124 10:21:27.075690 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:27 crc kubenswrapper[4944]: I1124 10:21:27.137241 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9vj97"] Nov 24 10:21:29 crc kubenswrapper[4944]: I1124 10:21:29.070441 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9vj97" podUID="80816194-484a-48d2-b749-b527341c4437" containerName="registry-server" containerID="cri-o://c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3" gracePeriod=2 Nov 24 10:21:29 crc kubenswrapper[4944]: I1124 10:21:29.489718 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:29 crc kubenswrapper[4944]: I1124 10:21:29.584696 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-utilities\") pod \"80816194-484a-48d2-b749-b527341c4437\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " Nov 24 10:21:29 crc kubenswrapper[4944]: I1124 10:21:29.584782 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgk4h\" (UniqueName: \"kubernetes.io/projected/80816194-484a-48d2-b749-b527341c4437-kube-api-access-qgk4h\") pod \"80816194-484a-48d2-b749-b527341c4437\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " Nov 24 10:21:29 crc kubenswrapper[4944]: I1124 10:21:29.584896 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-catalog-content\") pod \"80816194-484a-48d2-b749-b527341c4437\" (UID: \"80816194-484a-48d2-b749-b527341c4437\") " Nov 24 10:21:29 crc kubenswrapper[4944]: I1124 10:21:29.586706 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-utilities" (OuterVolumeSpecName: "utilities") pod "80816194-484a-48d2-b749-b527341c4437" (UID: "80816194-484a-48d2-b749-b527341c4437"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:21:29 crc kubenswrapper[4944]: I1124 10:21:29.589968 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80816194-484a-48d2-b749-b527341c4437-kube-api-access-qgk4h" (OuterVolumeSpecName: "kube-api-access-qgk4h") pod "80816194-484a-48d2-b749-b527341c4437" (UID: "80816194-484a-48d2-b749-b527341c4437"). InnerVolumeSpecName "kube-api-access-qgk4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:21:29 crc kubenswrapper[4944]: I1124 10:21:29.628701 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80816194-484a-48d2-b749-b527341c4437" (UID: "80816194-484a-48d2-b749-b527341c4437"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:21:29 crc kubenswrapper[4944]: I1124 10:21:29.686355 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:29 crc kubenswrapper[4944]: I1124 10:21:29.686383 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80816194-484a-48d2-b749-b527341c4437-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:29 crc kubenswrapper[4944]: I1124 10:21:29.686392 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgk4h\" (UniqueName: \"kubernetes.io/projected/80816194-484a-48d2-b749-b527341c4437-kube-api-access-qgk4h\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.080685 4944 generic.go:334] "Generic (PLEG): container finished" podID="80816194-484a-48d2-b749-b527341c4437" containerID="c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3" exitCode=0 Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.080798 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vj97" Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.081452 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vj97" event={"ID":"80816194-484a-48d2-b749-b527341c4437","Type":"ContainerDied","Data":"c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3"} Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.081488 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vj97" event={"ID":"80816194-484a-48d2-b749-b527341c4437","Type":"ContainerDied","Data":"c01b86d1d688f95b67e7d725240272bfe6605284af59acb745ef31604a2a9f4f"} Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.081506 4944 scope.go:117] "RemoveContainer" containerID="c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3" Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.109162 4944 scope.go:117] "RemoveContainer" containerID="e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a" Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.119994 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9vj97"] Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.130652 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9vj97"] Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.151619 4944 scope.go:117] "RemoveContainer" containerID="98f9cdfb08de6689174f6ce74169a359a31d5c8d124604d0d324321676be8df4" Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.184944 4944 scope.go:117] "RemoveContainer" containerID="c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3" Nov 24 10:21:30 crc kubenswrapper[4944]: E1124 10:21:30.185490 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3\": container with ID starting with c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3 not found: ID does not exist" containerID="c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3" Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.185543 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3"} err="failed to get container status \"c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3\": rpc error: code = NotFound desc = could not find container \"c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3\": container with ID starting with c7e3b2fa6028045ec743a20a7c1b1318923b64b238996977bb857cba4db965a3 not found: ID does not exist" Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.185579 4944 scope.go:117] "RemoveContainer" containerID="e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a" Nov 24 10:21:30 crc kubenswrapper[4944]: E1124 10:21:30.185950 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a\": container with ID starting with e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a not found: ID does not exist" containerID="e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a" Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.185993 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a"} err="failed to get container status \"e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a\": rpc error: code = NotFound desc = could not find container \"e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a\": container with ID starting with e664e1684931a37131d5aca0fe9f21cb5a5e2d8ec0acb0d56720f7ba839b110a not found: ID does not exist" Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.186015 4944 scope.go:117] "RemoveContainer" containerID="98f9cdfb08de6689174f6ce74169a359a31d5c8d124604d0d324321676be8df4" Nov 24 10:21:30 crc kubenswrapper[4944]: E1124 10:21:30.186329 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98f9cdfb08de6689174f6ce74169a359a31d5c8d124604d0d324321676be8df4\": container with ID starting with 98f9cdfb08de6689174f6ce74169a359a31d5c8d124604d0d324321676be8df4 not found: ID does not exist" containerID="98f9cdfb08de6689174f6ce74169a359a31d5c8d124604d0d324321676be8df4" Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.186370 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98f9cdfb08de6689174f6ce74169a359a31d5c8d124604d0d324321676be8df4"} err="failed to get container status \"98f9cdfb08de6689174f6ce74169a359a31d5c8d124604d0d324321676be8df4\": rpc error: code = NotFound desc = could not find container \"98f9cdfb08de6689174f6ce74169a359a31d5c8d124604d0d324321676be8df4\": container with ID starting with 98f9cdfb08de6689174f6ce74169a359a31d5c8d124604d0d324321676be8df4 not found: ID does not exist" Nov 24 10:21:30 crc kubenswrapper[4944]: I1124 10:21:30.286296 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80816194-484a-48d2-b749-b527341c4437" path="/var/lib/kubelet/pods/80816194-484a-48d2-b749-b527341c4437/volumes" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.030925 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8s9gc"] Nov 24 10:21:38 crc kubenswrapper[4944]: E1124 10:21:38.032445 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80816194-484a-48d2-b749-b527341c4437" containerName="registry-server" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.032460 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="80816194-484a-48d2-b749-b527341c4437" containerName="registry-server" Nov 24 10:21:38 crc kubenswrapper[4944]: E1124 10:21:38.032485 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6aaa08f-11fd-495a-88f4-e1684d0f229d" containerName="init" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.032491 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6aaa08f-11fd-495a-88f4-e1684d0f229d" containerName="init" Nov 24 10:21:38 crc kubenswrapper[4944]: E1124 10:21:38.032500 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80816194-484a-48d2-b749-b527341c4437" containerName="extract-utilities" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.032507 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="80816194-484a-48d2-b749-b527341c4437" containerName="extract-utilities" Nov 24 10:21:38 crc kubenswrapper[4944]: E1124 10:21:38.032526 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6aaa08f-11fd-495a-88f4-e1684d0f229d" containerName="dnsmasq-dns" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.032532 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6aaa08f-11fd-495a-88f4-e1684d0f229d" containerName="dnsmasq-dns" Nov 24 10:21:38 crc kubenswrapper[4944]: E1124 10:21:38.032544 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80816194-484a-48d2-b749-b527341c4437" containerName="extract-content" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.032549 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="80816194-484a-48d2-b749-b527341c4437" containerName="extract-content" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.032797 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="80816194-484a-48d2-b749-b527341c4437" containerName="registry-server" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.032824 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6aaa08f-11fd-495a-88f4-e1684d0f229d" containerName="dnsmasq-dns" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.035593 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.038809 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8s9gc"] Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.160097 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9n62\" (UniqueName: \"kubernetes.io/projected/35c1c85f-62fe-44d7-a990-d71cbf953ff6-kube-api-access-s9n62\") pod \"redhat-marketplace-8s9gc\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.160222 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-catalog-content\") pod \"redhat-marketplace-8s9gc\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.160317 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-utilities\") pod \"redhat-marketplace-8s9gc\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.261776 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-utilities\") pod \"redhat-marketplace-8s9gc\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.261856 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9n62\" (UniqueName: \"kubernetes.io/projected/35c1c85f-62fe-44d7-a990-d71cbf953ff6-kube-api-access-s9n62\") pod \"redhat-marketplace-8s9gc\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.261906 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-catalog-content\") pod \"redhat-marketplace-8s9gc\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.262399 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-catalog-content\") pod \"redhat-marketplace-8s9gc\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.262557 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-utilities\") pod \"redhat-marketplace-8s9gc\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.283143 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9n62\" (UniqueName: \"kubernetes.io/projected/35c1c85f-62fe-44d7-a990-d71cbf953ff6-kube-api-access-s9n62\") pod \"redhat-marketplace-8s9gc\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.359336 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:38 crc kubenswrapper[4944]: I1124 10:21:38.815185 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8s9gc"] Nov 24 10:21:39 crc kubenswrapper[4944]: I1124 10:21:39.156262 4944 generic.go:334] "Generic (PLEG): container finished" podID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" containerID="0ae1784bae5f3f95401964ea4f29eb4c88919b16988bff8212e12d08e436e3fa" exitCode=0 Nov 24 10:21:39 crc kubenswrapper[4944]: I1124 10:21:39.156372 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8s9gc" event={"ID":"35c1c85f-62fe-44d7-a990-d71cbf953ff6","Type":"ContainerDied","Data":"0ae1784bae5f3f95401964ea4f29eb4c88919b16988bff8212e12d08e436e3fa"} Nov 24 10:21:39 crc kubenswrapper[4944]: I1124 10:21:39.156631 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8s9gc" event={"ID":"35c1c85f-62fe-44d7-a990-d71cbf953ff6","Type":"ContainerStarted","Data":"62e344b3aa593bed6874363fa1c8e93a2dcc9fd41b1975b681ae74d01ad90ffc"} Nov 24 10:21:41 crc kubenswrapper[4944]: I1124 10:21:41.175741 4944 generic.go:334] "Generic (PLEG): container finished" podID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" containerID="745b158ea04b8bcf1acdda8f389c55804e1d03270d14ae6b027552b6f56fbfce" exitCode=0 Nov 24 10:21:41 crc kubenswrapper[4944]: I1124 10:21:41.175825 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8s9gc" event={"ID":"35c1c85f-62fe-44d7-a990-d71cbf953ff6","Type":"ContainerDied","Data":"745b158ea04b8bcf1acdda8f389c55804e1d03270d14ae6b027552b6f56fbfce"} Nov 24 10:21:42 crc kubenswrapper[4944]: I1124 10:21:42.193320 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8s9gc" event={"ID":"35c1c85f-62fe-44d7-a990-d71cbf953ff6","Type":"ContainerStarted","Data":"bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57"} Nov 24 10:21:42 crc kubenswrapper[4944]: I1124 10:21:42.216644 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8s9gc" podStartSLOduration=1.696121561 podStartE2EDuration="4.216621178s" podCreationTimestamp="2025-11-24 10:21:38 +0000 UTC" firstStartedPulling="2025-11-24 10:21:39.157447918 +0000 UTC m=+5359.691888380" lastFinishedPulling="2025-11-24 10:21:41.677947535 +0000 UTC m=+5362.212387997" observedRunningTime="2025-11-24 10:21:42.207190567 +0000 UTC m=+5362.741631039" watchObservedRunningTime="2025-11-24 10:21:42.216621178 +0000 UTC m=+5362.751061640" Nov 24 10:21:48 crc kubenswrapper[4944]: I1124 10:21:48.359760 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:48 crc kubenswrapper[4944]: I1124 10:21:48.360336 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:48 crc kubenswrapper[4944]: I1124 10:21:48.380741 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:48 crc kubenswrapper[4944]: I1124 10:21:48.411060 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:49 crc kubenswrapper[4944]: I1124 10:21:49.293794 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:49 crc kubenswrapper[4944]: I1124 10:21:49.382180 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8s9gc"] Nov 24 10:21:49 crc kubenswrapper[4944]: I1124 10:21:49.391668 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-774d97dd4-ks9d9" Nov 24 10:21:51 crc kubenswrapper[4944]: I1124 10:21:51.264243 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8s9gc" podUID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" containerName="registry-server" containerID="cri-o://bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57" gracePeriod=2 Nov 24 10:21:51 crc kubenswrapper[4944]: I1124 10:21:51.757484 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:51 crc kubenswrapper[4944]: I1124 10:21:51.831156 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-utilities\") pod \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " Nov 24 10:21:51 crc kubenswrapper[4944]: I1124 10:21:51.831314 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-catalog-content\") pod \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " Nov 24 10:21:51 crc kubenswrapper[4944]: I1124 10:21:51.831344 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9n62\" (UniqueName: \"kubernetes.io/projected/35c1c85f-62fe-44d7-a990-d71cbf953ff6-kube-api-access-s9n62\") pod \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\" (UID: \"35c1c85f-62fe-44d7-a990-d71cbf953ff6\") " Nov 24 10:21:51 crc kubenswrapper[4944]: I1124 10:21:51.832003 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-utilities" (OuterVolumeSpecName: "utilities") pod "35c1c85f-62fe-44d7-a990-d71cbf953ff6" (UID: "35c1c85f-62fe-44d7-a990-d71cbf953ff6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:21:51 crc kubenswrapper[4944]: I1124 10:21:51.837196 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35c1c85f-62fe-44d7-a990-d71cbf953ff6-kube-api-access-s9n62" (OuterVolumeSpecName: "kube-api-access-s9n62") pod "35c1c85f-62fe-44d7-a990-d71cbf953ff6" (UID: "35c1c85f-62fe-44d7-a990-d71cbf953ff6"). InnerVolumeSpecName "kube-api-access-s9n62". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:21:51 crc kubenswrapper[4944]: I1124 10:21:51.851170 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35c1c85f-62fe-44d7-a990-d71cbf953ff6" (UID: "35c1c85f-62fe-44d7-a990-d71cbf953ff6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:21:51 crc kubenswrapper[4944]: I1124 10:21:51.933156 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:51 crc kubenswrapper[4944]: I1124 10:21:51.933189 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9n62\" (UniqueName: \"kubernetes.io/projected/35c1c85f-62fe-44d7-a990-d71cbf953ff6-kube-api-access-s9n62\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:51 crc kubenswrapper[4944]: I1124 10:21:51.933200 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35c1c85f-62fe-44d7-a990-d71cbf953ff6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.273197 4944 generic.go:334] "Generic (PLEG): container finished" podID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" containerID="bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57" exitCode=0 Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.273610 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8s9gc" event={"ID":"35c1c85f-62fe-44d7-a990-d71cbf953ff6","Type":"ContainerDied","Data":"bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57"} Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.273643 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8s9gc" event={"ID":"35c1c85f-62fe-44d7-a990-d71cbf953ff6","Type":"ContainerDied","Data":"62e344b3aa593bed6874363fa1c8e93a2dcc9fd41b1975b681ae74d01ad90ffc"} Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.273664 4944 scope.go:117] "RemoveContainer" containerID="bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57" Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.273816 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8s9gc" Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.308437 4944 scope.go:117] "RemoveContainer" containerID="745b158ea04b8bcf1acdda8f389c55804e1d03270d14ae6b027552b6f56fbfce" Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.323059 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8s9gc"] Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.331083 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8s9gc"] Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.341409 4944 scope.go:117] "RemoveContainer" containerID="0ae1784bae5f3f95401964ea4f29eb4c88919b16988bff8212e12d08e436e3fa" Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.389887 4944 scope.go:117] "RemoveContainer" containerID="bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57" Nov 24 10:21:52 crc kubenswrapper[4944]: E1124 10:21:52.390489 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57\": container with ID starting with bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57 not found: ID does not exist" containerID="bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57" Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.390526 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57"} err="failed to get container status \"bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57\": rpc error: code = NotFound desc = could not find container \"bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57\": container with ID starting with bdfd15cd2258752b43d31bc4f0b66969eab531d4f1f016d5207a9ee017345b57 not found: ID does not exist" Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.390550 4944 scope.go:117] "RemoveContainer" containerID="745b158ea04b8bcf1acdda8f389c55804e1d03270d14ae6b027552b6f56fbfce" Nov 24 10:21:52 crc kubenswrapper[4944]: E1124 10:21:52.391111 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"745b158ea04b8bcf1acdda8f389c55804e1d03270d14ae6b027552b6f56fbfce\": container with ID starting with 745b158ea04b8bcf1acdda8f389c55804e1d03270d14ae6b027552b6f56fbfce not found: ID does not exist" containerID="745b158ea04b8bcf1acdda8f389c55804e1d03270d14ae6b027552b6f56fbfce" Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.391141 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"745b158ea04b8bcf1acdda8f389c55804e1d03270d14ae6b027552b6f56fbfce"} err="failed to get container status \"745b158ea04b8bcf1acdda8f389c55804e1d03270d14ae6b027552b6f56fbfce\": rpc error: code = NotFound desc = could not find container \"745b158ea04b8bcf1acdda8f389c55804e1d03270d14ae6b027552b6f56fbfce\": container with ID starting with 745b158ea04b8bcf1acdda8f389c55804e1d03270d14ae6b027552b6f56fbfce not found: ID does not exist" Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.391161 4944 scope.go:117] "RemoveContainer" containerID="0ae1784bae5f3f95401964ea4f29eb4c88919b16988bff8212e12d08e436e3fa" Nov 24 10:21:52 crc kubenswrapper[4944]: E1124 10:21:52.391724 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ae1784bae5f3f95401964ea4f29eb4c88919b16988bff8212e12d08e436e3fa\": container with ID starting with 0ae1784bae5f3f95401964ea4f29eb4c88919b16988bff8212e12d08e436e3fa not found: ID does not exist" containerID="0ae1784bae5f3f95401964ea4f29eb4c88919b16988bff8212e12d08e436e3fa" Nov 24 10:21:52 crc kubenswrapper[4944]: I1124 10:21:52.391783 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ae1784bae5f3f95401964ea4f29eb4c88919b16988bff8212e12d08e436e3fa"} err="failed to get container status \"0ae1784bae5f3f95401964ea4f29eb4c88919b16988bff8212e12d08e436e3fa\": rpc error: code = NotFound desc = could not find container \"0ae1784bae5f3f95401964ea4f29eb4c88919b16988bff8212e12d08e436e3fa\": container with ID starting with 0ae1784bae5f3f95401964ea4f29eb4c88919b16988bff8212e12d08e436e3fa not found: ID does not exist" Nov 24 10:21:54 crc kubenswrapper[4944]: I1124 10:21:54.289737 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" path="/var/lib/kubelet/pods/35c1c85f-62fe-44d7-a990-d71cbf953ff6/volumes" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.343610 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8mh4m"] Nov 24 10:22:08 crc kubenswrapper[4944]: E1124 10:22:08.344559 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" containerName="registry-server" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.344572 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" containerName="registry-server" Nov 24 10:22:08 crc kubenswrapper[4944]: E1124 10:22:08.344583 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" containerName="extract-utilities" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.344589 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" containerName="extract-utilities" Nov 24 10:22:08 crc kubenswrapper[4944]: E1124 10:22:08.344619 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" containerName="extract-content" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.344625 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" containerName="extract-content" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.344791 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="35c1c85f-62fe-44d7-a990-d71cbf953ff6" containerName="registry-server" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.345356 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8mh4m" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.354140 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8mh4m"] Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.411030 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aae14ddf-150b-4d93-9632-b6f3dff92eea-operator-scripts\") pod \"nova-api-db-create-8mh4m\" (UID: \"aae14ddf-150b-4d93-9632-b6f3dff92eea\") " pod="openstack/nova-api-db-create-8mh4m" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.411211 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h89d2\" (UniqueName: \"kubernetes.io/projected/aae14ddf-150b-4d93-9632-b6f3dff92eea-kube-api-access-h89d2\") pod \"nova-api-db-create-8mh4m\" (UID: \"aae14ddf-150b-4d93-9632-b6f3dff92eea\") " pod="openstack/nova-api-db-create-8mh4m" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.434949 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-mvbzx"] Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.436070 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mvbzx" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.443499 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-mvbzx"] Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.511946 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bf65b97-ae89-462d-9bde-d694c17f8c46-operator-scripts\") pod \"nova-cell0-db-create-mvbzx\" (UID: \"6bf65b97-ae89-462d-9bde-d694c17f8c46\") " pod="openstack/nova-cell0-db-create-mvbzx" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.512006 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h89d2\" (UniqueName: \"kubernetes.io/projected/aae14ddf-150b-4d93-9632-b6f3dff92eea-kube-api-access-h89d2\") pod \"nova-api-db-create-8mh4m\" (UID: \"aae14ddf-150b-4d93-9632-b6f3dff92eea\") " pod="openstack/nova-api-db-create-8mh4m" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.512038 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjlqd\" (UniqueName: \"kubernetes.io/projected/6bf65b97-ae89-462d-9bde-d694c17f8c46-kube-api-access-xjlqd\") pod \"nova-cell0-db-create-mvbzx\" (UID: \"6bf65b97-ae89-462d-9bde-d694c17f8c46\") " pod="openstack/nova-cell0-db-create-mvbzx" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.512105 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aae14ddf-150b-4d93-9632-b6f3dff92eea-operator-scripts\") pod \"nova-api-db-create-8mh4m\" (UID: \"aae14ddf-150b-4d93-9632-b6f3dff92eea\") " pod="openstack/nova-api-db-create-8mh4m" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.512873 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aae14ddf-150b-4d93-9632-b6f3dff92eea-operator-scripts\") pod \"nova-api-db-create-8mh4m\" (UID: \"aae14ddf-150b-4d93-9632-b6f3dff92eea\") " pod="openstack/nova-api-db-create-8mh4m" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.538784 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h89d2\" (UniqueName: \"kubernetes.io/projected/aae14ddf-150b-4d93-9632-b6f3dff92eea-kube-api-access-h89d2\") pod \"nova-api-db-create-8mh4m\" (UID: \"aae14ddf-150b-4d93-9632-b6f3dff92eea\") " pod="openstack/nova-api-db-create-8mh4m" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.545337 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-9dab-account-create-r8kk5"] Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.546698 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9dab-account-create-r8kk5" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.549452 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.553871 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-x9f7d"] Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.555194 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x9f7d" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.571272 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-x9f7d"] Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.580301 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-9dab-account-create-r8kk5"] Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.613708 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-operator-scripts\") pod \"nova-cell1-db-create-x9f7d\" (UID: \"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c\") " pod="openstack/nova-cell1-db-create-x9f7d" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.614123 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bf65b97-ae89-462d-9bde-d694c17f8c46-operator-scripts\") pod \"nova-cell0-db-create-mvbzx\" (UID: \"6bf65b97-ae89-462d-9bde-d694c17f8c46\") " pod="openstack/nova-cell0-db-create-mvbzx" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.614162 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8269344b-cb43-4154-8550-84147c6deb7d-operator-scripts\") pod \"nova-api-9dab-account-create-r8kk5\" (UID: \"8269344b-cb43-4154-8550-84147c6deb7d\") " pod="openstack/nova-api-9dab-account-create-r8kk5" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.614198 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjlqd\" (UniqueName: \"kubernetes.io/projected/6bf65b97-ae89-462d-9bde-d694c17f8c46-kube-api-access-xjlqd\") pod \"nova-cell0-db-create-mvbzx\" (UID: \"6bf65b97-ae89-462d-9bde-d694c17f8c46\") " pod="openstack/nova-cell0-db-create-mvbzx" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.614222 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw82v\" (UniqueName: \"kubernetes.io/projected/8269344b-cb43-4154-8550-84147c6deb7d-kube-api-access-tw82v\") pod \"nova-api-9dab-account-create-r8kk5\" (UID: \"8269344b-cb43-4154-8550-84147c6deb7d\") " pod="openstack/nova-api-9dab-account-create-r8kk5" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.614316 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r984\" (UniqueName: \"kubernetes.io/projected/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-kube-api-access-9r984\") pod \"nova-cell1-db-create-x9f7d\" (UID: \"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c\") " pod="openstack/nova-cell1-db-create-x9f7d" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.614924 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bf65b97-ae89-462d-9bde-d694c17f8c46-operator-scripts\") pod \"nova-cell0-db-create-mvbzx\" (UID: \"6bf65b97-ae89-462d-9bde-d694c17f8c46\") " pod="openstack/nova-cell0-db-create-mvbzx" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.632797 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjlqd\" (UniqueName: \"kubernetes.io/projected/6bf65b97-ae89-462d-9bde-d694c17f8c46-kube-api-access-xjlqd\") pod \"nova-cell0-db-create-mvbzx\" (UID: \"6bf65b97-ae89-462d-9bde-d694c17f8c46\") " pod="openstack/nova-cell0-db-create-mvbzx" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.670925 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8mh4m" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.715232 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r984\" (UniqueName: \"kubernetes.io/projected/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-kube-api-access-9r984\") pod \"nova-cell1-db-create-x9f7d\" (UID: \"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c\") " pod="openstack/nova-cell1-db-create-x9f7d" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.715310 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-operator-scripts\") pod \"nova-cell1-db-create-x9f7d\" (UID: \"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c\") " pod="openstack/nova-cell1-db-create-x9f7d" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.715384 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8269344b-cb43-4154-8550-84147c6deb7d-operator-scripts\") pod \"nova-api-9dab-account-create-r8kk5\" (UID: \"8269344b-cb43-4154-8550-84147c6deb7d\") " pod="openstack/nova-api-9dab-account-create-r8kk5" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.715432 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw82v\" (UniqueName: \"kubernetes.io/projected/8269344b-cb43-4154-8550-84147c6deb7d-kube-api-access-tw82v\") pod \"nova-api-9dab-account-create-r8kk5\" (UID: \"8269344b-cb43-4154-8550-84147c6deb7d\") " pod="openstack/nova-api-9dab-account-create-r8kk5" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.716218 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8269344b-cb43-4154-8550-84147c6deb7d-operator-scripts\") pod \"nova-api-9dab-account-create-r8kk5\" (UID: \"8269344b-cb43-4154-8550-84147c6deb7d\") " pod="openstack/nova-api-9dab-account-create-r8kk5" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.717845 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-operator-scripts\") pod \"nova-cell1-db-create-x9f7d\" (UID: \"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c\") " pod="openstack/nova-cell1-db-create-x9f7d" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.735029 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r984\" (UniqueName: \"kubernetes.io/projected/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-kube-api-access-9r984\") pod \"nova-cell1-db-create-x9f7d\" (UID: \"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c\") " pod="openstack/nova-cell1-db-create-x9f7d" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.738763 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw82v\" (UniqueName: \"kubernetes.io/projected/8269344b-cb43-4154-8550-84147c6deb7d-kube-api-access-tw82v\") pod \"nova-api-9dab-account-create-r8kk5\" (UID: \"8269344b-cb43-4154-8550-84147c6deb7d\") " pod="openstack/nova-api-9dab-account-create-r8kk5" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.744012 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-23ab-account-create-zv4cc"] Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.745255 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-23ab-account-create-zv4cc" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.750302 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mvbzx" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.755270 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.755558 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-23ab-account-create-zv4cc"] Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.854717 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-023c-account-create-9xqh8"] Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.858142 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-023c-account-create-9xqh8" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.861639 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.862976 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-023c-account-create-9xqh8"] Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.888145 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9dab-account-create-r8kk5" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.896520 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x9f7d" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.919314 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24c9aee6-8750-4862-8030-cc249e7464b2-operator-scripts\") pod \"nova-cell0-23ab-account-create-zv4cc\" (UID: \"24c9aee6-8750-4862-8030-cc249e7464b2\") " pod="openstack/nova-cell0-23ab-account-create-zv4cc" Nov 24 10:22:08 crc kubenswrapper[4944]: I1124 10:22:08.919664 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h9nz\" (UniqueName: \"kubernetes.io/projected/24c9aee6-8750-4862-8030-cc249e7464b2-kube-api-access-9h9nz\") pod \"nova-cell0-23ab-account-create-zv4cc\" (UID: \"24c9aee6-8750-4862-8030-cc249e7464b2\") " pod="openstack/nova-cell0-23ab-account-create-zv4cc" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.022015 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4490175c-09f1-4ba1-8842-5a9a576ffce1-operator-scripts\") pod \"nova-cell1-023c-account-create-9xqh8\" (UID: \"4490175c-09f1-4ba1-8842-5a9a576ffce1\") " pod="openstack/nova-cell1-023c-account-create-9xqh8" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.022256 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h9nz\" (UniqueName: \"kubernetes.io/projected/24c9aee6-8750-4862-8030-cc249e7464b2-kube-api-access-9h9nz\") pod \"nova-cell0-23ab-account-create-zv4cc\" (UID: \"24c9aee6-8750-4862-8030-cc249e7464b2\") " pod="openstack/nova-cell0-23ab-account-create-zv4cc" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.022430 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24c9aee6-8750-4862-8030-cc249e7464b2-operator-scripts\") pod \"nova-cell0-23ab-account-create-zv4cc\" (UID: \"24c9aee6-8750-4862-8030-cc249e7464b2\") " pod="openstack/nova-cell0-23ab-account-create-zv4cc" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.022569 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mztp5\" (UniqueName: \"kubernetes.io/projected/4490175c-09f1-4ba1-8842-5a9a576ffce1-kube-api-access-mztp5\") pod \"nova-cell1-023c-account-create-9xqh8\" (UID: \"4490175c-09f1-4ba1-8842-5a9a576ffce1\") " pod="openstack/nova-cell1-023c-account-create-9xqh8" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.023387 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24c9aee6-8750-4862-8030-cc249e7464b2-operator-scripts\") pod \"nova-cell0-23ab-account-create-zv4cc\" (UID: \"24c9aee6-8750-4862-8030-cc249e7464b2\") " pod="openstack/nova-cell0-23ab-account-create-zv4cc" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.040595 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h9nz\" (UniqueName: \"kubernetes.io/projected/24c9aee6-8750-4862-8030-cc249e7464b2-kube-api-access-9h9nz\") pod \"nova-cell0-23ab-account-create-zv4cc\" (UID: \"24c9aee6-8750-4862-8030-cc249e7464b2\") " pod="openstack/nova-cell0-23ab-account-create-zv4cc" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.125269 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mztp5\" (UniqueName: \"kubernetes.io/projected/4490175c-09f1-4ba1-8842-5a9a576ffce1-kube-api-access-mztp5\") pod \"nova-cell1-023c-account-create-9xqh8\" (UID: \"4490175c-09f1-4ba1-8842-5a9a576ffce1\") " pod="openstack/nova-cell1-023c-account-create-9xqh8" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.125337 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4490175c-09f1-4ba1-8842-5a9a576ffce1-operator-scripts\") pod \"nova-cell1-023c-account-create-9xqh8\" (UID: \"4490175c-09f1-4ba1-8842-5a9a576ffce1\") " pod="openstack/nova-cell1-023c-account-create-9xqh8" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.126419 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4490175c-09f1-4ba1-8842-5a9a576ffce1-operator-scripts\") pod \"nova-cell1-023c-account-create-9xqh8\" (UID: \"4490175c-09f1-4ba1-8842-5a9a576ffce1\") " pod="openstack/nova-cell1-023c-account-create-9xqh8" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.143647 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mztp5\" (UniqueName: \"kubernetes.io/projected/4490175c-09f1-4ba1-8842-5a9a576ffce1-kube-api-access-mztp5\") pod \"nova-cell1-023c-account-create-9xqh8\" (UID: \"4490175c-09f1-4ba1-8842-5a9a576ffce1\") " pod="openstack/nova-cell1-023c-account-create-9xqh8" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.157598 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-23ab-account-create-zv4cc" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.186972 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-023c-account-create-9xqh8" Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.240615 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8mh4m"] Nov 24 10:22:09 crc kubenswrapper[4944]: W1124 10:22:09.256620 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaae14ddf_150b_4d93_9632_b6f3dff92eea.slice/crio-6b968c2fafda0f3c2a9de79acdf4617ff46e29de9f599588259580c8a9996b67 WatchSource:0}: Error finding container 6b968c2fafda0f3c2a9de79acdf4617ff46e29de9f599588259580c8a9996b67: Status 404 returned error can't find the container with id 6b968c2fafda0f3c2a9de79acdf4617ff46e29de9f599588259580c8a9996b67 Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.358228 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-mvbzx"] Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.434322 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-9dab-account-create-r8kk5"] Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.436400 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-mvbzx" event={"ID":"6bf65b97-ae89-462d-9bde-d694c17f8c46","Type":"ContainerStarted","Data":"40811674fd1538bc37ffd038be9d10166818fbf10b862128c6f152b556e69578"} Nov 24 10:22:09 crc kubenswrapper[4944]: W1124 10:22:09.436848 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8269344b_cb43_4154_8550_84147c6deb7d.slice/crio-f30eefb54f57f2f596c0e5c97a727327fac63d7b96a705a4eab5fe8eb622b7c8 WatchSource:0}: Error finding container f30eefb54f57f2f596c0e5c97a727327fac63d7b96a705a4eab5fe8eb622b7c8: Status 404 returned error can't find the container with id f30eefb54f57f2f596c0e5c97a727327fac63d7b96a705a4eab5fe8eb622b7c8 Nov 24 10:22:09 crc kubenswrapper[4944]: W1124 10:22:09.439360 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfd0d9ab_a8b9_4104_bb7b_1b1bac2bad2c.slice/crio-10cdd6b66f1f91bec26b69f69cff420c48f20a8d5ec4601cadf6935ef446df9d WatchSource:0}: Error finding container 10cdd6b66f1f91bec26b69f69cff420c48f20a8d5ec4601cadf6935ef446df9d: Status 404 returned error can't find the container with id 10cdd6b66f1f91bec26b69f69cff420c48f20a8d5ec4601cadf6935ef446df9d Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.442484 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-x9f7d"] Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.443636 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8mh4m" event={"ID":"aae14ddf-150b-4d93-9632-b6f3dff92eea","Type":"ContainerStarted","Data":"6b968c2fafda0f3c2a9de79acdf4617ff46e29de9f599588259580c8a9996b67"} Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.627025 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-23ab-account-create-zv4cc"] Nov 24 10:22:09 crc kubenswrapper[4944]: I1124 10:22:09.758727 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-023c-account-create-9xqh8"] Nov 24 10:22:09 crc kubenswrapper[4944]: W1124 10:22:09.759731 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4490175c_09f1_4ba1_8842_5a9a576ffce1.slice/crio-2ce6ca15896d7af94e700d1a9562ad340e27c86d1adc6b268e42577165a68019 WatchSource:0}: Error finding container 2ce6ca15896d7af94e700d1a9562ad340e27c86d1adc6b268e42577165a68019: Status 404 returned error can't find the container with id 2ce6ca15896d7af94e700d1a9562ad340e27c86d1adc6b268e42577165a68019 Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.454290 4944 generic.go:334] "Generic (PLEG): container finished" podID="aae14ddf-150b-4d93-9632-b6f3dff92eea" containerID="2b1490de7e5d194d5a177b2bcb42fd34098c29b5bf9de2e5a82c0b6430805910" exitCode=0 Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.454387 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8mh4m" event={"ID":"aae14ddf-150b-4d93-9632-b6f3dff92eea","Type":"ContainerDied","Data":"2b1490de7e5d194d5a177b2bcb42fd34098c29b5bf9de2e5a82c0b6430805910"} Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.458160 4944 generic.go:334] "Generic (PLEG): container finished" podID="24c9aee6-8750-4862-8030-cc249e7464b2" containerID="5d7739bd33ffb92b9bbf110c7a163cd4f76dd6925a8f1631772040af87b2bfd3" exitCode=0 Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.458232 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-23ab-account-create-zv4cc" event={"ID":"24c9aee6-8750-4862-8030-cc249e7464b2","Type":"ContainerDied","Data":"5d7739bd33ffb92b9bbf110c7a163cd4f76dd6925a8f1631772040af87b2bfd3"} Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.458360 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-23ab-account-create-zv4cc" event={"ID":"24c9aee6-8750-4862-8030-cc249e7464b2","Type":"ContainerStarted","Data":"b9de35395a1f76f84624c0b5e06560df403a390da20e88a2ed1ad6cb6ccddb31"} Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.459901 4944 generic.go:334] "Generic (PLEG): container finished" podID="6bf65b97-ae89-462d-9bde-d694c17f8c46" containerID="4269de9f24f4eb215c2e66c0fc50eeb78e34b2d967b36452fea0d2f63b6121f1" exitCode=0 Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.459976 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-mvbzx" event={"ID":"6bf65b97-ae89-462d-9bde-d694c17f8c46","Type":"ContainerDied","Data":"4269de9f24f4eb215c2e66c0fc50eeb78e34b2d967b36452fea0d2f63b6121f1"} Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.462576 4944 generic.go:334] "Generic (PLEG): container finished" podID="dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c" containerID="002f641c5545a01ed995891d667a5662ec851ca2b44e2266a4daef8073bfc478" exitCode=0 Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.462644 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-x9f7d" event={"ID":"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c","Type":"ContainerDied","Data":"002f641c5545a01ed995891d667a5662ec851ca2b44e2266a4daef8073bfc478"} Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.462672 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-x9f7d" event={"ID":"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c","Type":"ContainerStarted","Data":"10cdd6b66f1f91bec26b69f69cff420c48f20a8d5ec4601cadf6935ef446df9d"} Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.465560 4944 generic.go:334] "Generic (PLEG): container finished" podID="4490175c-09f1-4ba1-8842-5a9a576ffce1" containerID="083d4742ebc41b7ca3957c3ef4d36ce42a9b5f560f247bb1dbc1572aba343811" exitCode=0 Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.465630 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-023c-account-create-9xqh8" event={"ID":"4490175c-09f1-4ba1-8842-5a9a576ffce1","Type":"ContainerDied","Data":"083d4742ebc41b7ca3957c3ef4d36ce42a9b5f560f247bb1dbc1572aba343811"} Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.465649 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-023c-account-create-9xqh8" event={"ID":"4490175c-09f1-4ba1-8842-5a9a576ffce1","Type":"ContainerStarted","Data":"2ce6ca15896d7af94e700d1a9562ad340e27c86d1adc6b268e42577165a68019"} Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.470375 4944 generic.go:334] "Generic (PLEG): container finished" podID="8269344b-cb43-4154-8550-84147c6deb7d" containerID="78af2b150e96a2d80d31547eed468b9b38c6965eab76d0cc7e2e8cb47334d259" exitCode=0 Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.470437 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9dab-account-create-r8kk5" event={"ID":"8269344b-cb43-4154-8550-84147c6deb7d","Type":"ContainerDied","Data":"78af2b150e96a2d80d31547eed468b9b38c6965eab76d0cc7e2e8cb47334d259"} Nov 24 10:22:10 crc kubenswrapper[4944]: I1124 10:22:10.470473 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9dab-account-create-r8kk5" event={"ID":"8269344b-cb43-4154-8550-84147c6deb7d","Type":"ContainerStarted","Data":"f30eefb54f57f2f596c0e5c97a727327fac63d7b96a705a4eab5fe8eb622b7c8"} Nov 24 10:22:11 crc kubenswrapper[4944]: I1124 10:22:11.833639 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x9f7d" Nov 24 10:22:11 crc kubenswrapper[4944]: I1124 10:22:11.980588 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-operator-scripts\") pod \"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c\" (UID: \"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c\") " Nov 24 10:22:11 crc kubenswrapper[4944]: I1124 10:22:11.980928 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r984\" (UniqueName: \"kubernetes.io/projected/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-kube-api-access-9r984\") pod \"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c\" (UID: \"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c\") " Nov 24 10:22:11 crc kubenswrapper[4944]: I1124 10:22:11.981722 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c" (UID: "dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:22:11 crc kubenswrapper[4944]: I1124 10:22:11.981888 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:11 crc kubenswrapper[4944]: I1124 10:22:11.987002 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-kube-api-access-9r984" (OuterVolumeSpecName: "kube-api-access-9r984") pod "dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c" (UID: "dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c"). InnerVolumeSpecName "kube-api-access-9r984". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.061990 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-23ab-account-create-zv4cc" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.067750 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8mh4m" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.080877 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mvbzx" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.086612 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r984\" (UniqueName: \"kubernetes.io/projected/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c-kube-api-access-9r984\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.093192 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9dab-account-create-r8kk5" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.103421 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-023c-account-create-9xqh8" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.187504 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw82v\" (UniqueName: \"kubernetes.io/projected/8269344b-cb43-4154-8550-84147c6deb7d-kube-api-access-tw82v\") pod \"8269344b-cb43-4154-8550-84147c6deb7d\" (UID: \"8269344b-cb43-4154-8550-84147c6deb7d\") " Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.187618 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bf65b97-ae89-462d-9bde-d694c17f8c46-operator-scripts\") pod \"6bf65b97-ae89-462d-9bde-d694c17f8c46\" (UID: \"6bf65b97-ae89-462d-9bde-d694c17f8c46\") " Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.187652 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8269344b-cb43-4154-8550-84147c6deb7d-operator-scripts\") pod \"8269344b-cb43-4154-8550-84147c6deb7d\" (UID: \"8269344b-cb43-4154-8550-84147c6deb7d\") " Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.187716 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24c9aee6-8750-4862-8030-cc249e7464b2-operator-scripts\") pod \"24c9aee6-8750-4862-8030-cc249e7464b2\" (UID: \"24c9aee6-8750-4862-8030-cc249e7464b2\") " Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.187762 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h89d2\" (UniqueName: \"kubernetes.io/projected/aae14ddf-150b-4d93-9632-b6f3dff92eea-kube-api-access-h89d2\") pod \"aae14ddf-150b-4d93-9632-b6f3dff92eea\" (UID: \"aae14ddf-150b-4d93-9632-b6f3dff92eea\") " Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.187894 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aae14ddf-150b-4d93-9632-b6f3dff92eea-operator-scripts\") pod \"aae14ddf-150b-4d93-9632-b6f3dff92eea\" (UID: \"aae14ddf-150b-4d93-9632-b6f3dff92eea\") " Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.188130 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bf65b97-ae89-462d-9bde-d694c17f8c46-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6bf65b97-ae89-462d-9bde-d694c17f8c46" (UID: "6bf65b97-ae89-462d-9bde-d694c17f8c46"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.188296 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8269344b-cb43-4154-8550-84147c6deb7d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8269344b-cb43-4154-8550-84147c6deb7d" (UID: "8269344b-cb43-4154-8550-84147c6deb7d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.188341 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24c9aee6-8750-4862-8030-cc249e7464b2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "24c9aee6-8750-4862-8030-cc249e7464b2" (UID: "24c9aee6-8750-4862-8030-cc249e7464b2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.188437 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mztp5\" (UniqueName: \"kubernetes.io/projected/4490175c-09f1-4ba1-8842-5a9a576ffce1-kube-api-access-mztp5\") pod \"4490175c-09f1-4ba1-8842-5a9a576ffce1\" (UID: \"4490175c-09f1-4ba1-8842-5a9a576ffce1\") " Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.188448 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae14ddf-150b-4d93-9632-b6f3dff92eea-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aae14ddf-150b-4d93-9632-b6f3dff92eea" (UID: "aae14ddf-150b-4d93-9632-b6f3dff92eea"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.188537 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4490175c-09f1-4ba1-8842-5a9a576ffce1-operator-scripts\") pod \"4490175c-09f1-4ba1-8842-5a9a576ffce1\" (UID: \"4490175c-09f1-4ba1-8842-5a9a576ffce1\") " Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.188623 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h9nz\" (UniqueName: \"kubernetes.io/projected/24c9aee6-8750-4862-8030-cc249e7464b2-kube-api-access-9h9nz\") pod \"24c9aee6-8750-4862-8030-cc249e7464b2\" (UID: \"24c9aee6-8750-4862-8030-cc249e7464b2\") " Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.188681 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjlqd\" (UniqueName: \"kubernetes.io/projected/6bf65b97-ae89-462d-9bde-d694c17f8c46-kube-api-access-xjlqd\") pod \"6bf65b97-ae89-462d-9bde-d694c17f8c46\" (UID: \"6bf65b97-ae89-462d-9bde-d694c17f8c46\") " Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.188996 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4490175c-09f1-4ba1-8842-5a9a576ffce1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4490175c-09f1-4ba1-8842-5a9a576ffce1" (UID: "4490175c-09f1-4ba1-8842-5a9a576ffce1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.189370 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bf65b97-ae89-462d-9bde-d694c17f8c46-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.189396 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8269344b-cb43-4154-8550-84147c6deb7d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.189409 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24c9aee6-8750-4862-8030-cc249e7464b2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.189420 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aae14ddf-150b-4d93-9632-b6f3dff92eea-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.189431 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4490175c-09f1-4ba1-8842-5a9a576ffce1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.190638 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8269344b-cb43-4154-8550-84147c6deb7d-kube-api-access-tw82v" (OuterVolumeSpecName: "kube-api-access-tw82v") pod "8269344b-cb43-4154-8550-84147c6deb7d" (UID: "8269344b-cb43-4154-8550-84147c6deb7d"). InnerVolumeSpecName "kube-api-access-tw82v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.191200 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bf65b97-ae89-462d-9bde-d694c17f8c46-kube-api-access-xjlqd" (OuterVolumeSpecName: "kube-api-access-xjlqd") pod "6bf65b97-ae89-462d-9bde-d694c17f8c46" (UID: "6bf65b97-ae89-462d-9bde-d694c17f8c46"). InnerVolumeSpecName "kube-api-access-xjlqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.191339 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4490175c-09f1-4ba1-8842-5a9a576ffce1-kube-api-access-mztp5" (OuterVolumeSpecName: "kube-api-access-mztp5") pod "4490175c-09f1-4ba1-8842-5a9a576ffce1" (UID: "4490175c-09f1-4ba1-8842-5a9a576ffce1"). InnerVolumeSpecName "kube-api-access-mztp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.191473 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24c9aee6-8750-4862-8030-cc249e7464b2-kube-api-access-9h9nz" (OuterVolumeSpecName: "kube-api-access-9h9nz") pod "24c9aee6-8750-4862-8030-cc249e7464b2" (UID: "24c9aee6-8750-4862-8030-cc249e7464b2"). InnerVolumeSpecName "kube-api-access-9h9nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.191588 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae14ddf-150b-4d93-9632-b6f3dff92eea-kube-api-access-h89d2" (OuterVolumeSpecName: "kube-api-access-h89d2") pod "aae14ddf-150b-4d93-9632-b6f3dff92eea" (UID: "aae14ddf-150b-4d93-9632-b6f3dff92eea"). InnerVolumeSpecName "kube-api-access-h89d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.290723 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw82v\" (UniqueName: \"kubernetes.io/projected/8269344b-cb43-4154-8550-84147c6deb7d-kube-api-access-tw82v\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.290765 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h89d2\" (UniqueName: \"kubernetes.io/projected/aae14ddf-150b-4d93-9632-b6f3dff92eea-kube-api-access-h89d2\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.290777 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mztp5\" (UniqueName: \"kubernetes.io/projected/4490175c-09f1-4ba1-8842-5a9a576ffce1-kube-api-access-mztp5\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.290789 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h9nz\" (UniqueName: \"kubernetes.io/projected/24c9aee6-8750-4862-8030-cc249e7464b2-kube-api-access-9h9nz\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.290802 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjlqd\" (UniqueName: \"kubernetes.io/projected/6bf65b97-ae89-462d-9bde-d694c17f8c46-kube-api-access-xjlqd\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.496516 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8mh4m" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.496514 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8mh4m" event={"ID":"aae14ddf-150b-4d93-9632-b6f3dff92eea","Type":"ContainerDied","Data":"6b968c2fafda0f3c2a9de79acdf4617ff46e29de9f599588259580c8a9996b67"} Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.496577 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b968c2fafda0f3c2a9de79acdf4617ff46e29de9f599588259580c8a9996b67" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.500490 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-23ab-account-create-zv4cc" event={"ID":"24c9aee6-8750-4862-8030-cc249e7464b2","Type":"ContainerDied","Data":"b9de35395a1f76f84624c0b5e06560df403a390da20e88a2ed1ad6cb6ccddb31"} Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.500512 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9de35395a1f76f84624c0b5e06560df403a390da20e88a2ed1ad6cb6ccddb31" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.500575 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-23ab-account-create-zv4cc" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.502814 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-mvbzx" event={"ID":"6bf65b97-ae89-462d-9bde-d694c17f8c46","Type":"ContainerDied","Data":"40811674fd1538bc37ffd038be9d10166818fbf10b862128c6f152b556e69578"} Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.502849 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40811674fd1538bc37ffd038be9d10166818fbf10b862128c6f152b556e69578" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.502885 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-mvbzx" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.504386 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-x9f7d" event={"ID":"dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c","Type":"ContainerDied","Data":"10cdd6b66f1f91bec26b69f69cff420c48f20a8d5ec4601cadf6935ef446df9d"} Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.504418 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x9f7d" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.504426 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10cdd6b66f1f91bec26b69f69cff420c48f20a8d5ec4601cadf6935ef446df9d" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.507737 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-023c-account-create-9xqh8" event={"ID":"4490175c-09f1-4ba1-8842-5a9a576ffce1","Type":"ContainerDied","Data":"2ce6ca15896d7af94e700d1a9562ad340e27c86d1adc6b268e42577165a68019"} Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.507761 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ce6ca15896d7af94e700d1a9562ad340e27c86d1adc6b268e42577165a68019" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.507796 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-023c-account-create-9xqh8" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.509568 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9dab-account-create-r8kk5" event={"ID":"8269344b-cb43-4154-8550-84147c6deb7d","Type":"ContainerDied","Data":"f30eefb54f57f2f596c0e5c97a727327fac63d7b96a705a4eab5fe8eb622b7c8"} Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.509609 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f30eefb54f57f2f596c0e5c97a727327fac63d7b96a705a4eab5fe8eb622b7c8" Nov 24 10:22:12 crc kubenswrapper[4944]: I1124 10:22:12.509616 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9dab-account-create-r8kk5" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.921612 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l44vr"] Nov 24 10:22:13 crc kubenswrapper[4944]: E1124 10:22:13.922352 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4490175c-09f1-4ba1-8842-5a9a576ffce1" containerName="mariadb-account-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922369 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4490175c-09f1-4ba1-8842-5a9a576ffce1" containerName="mariadb-account-create" Nov 24 10:22:13 crc kubenswrapper[4944]: E1124 10:22:13.922387 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24c9aee6-8750-4862-8030-cc249e7464b2" containerName="mariadb-account-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922394 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="24c9aee6-8750-4862-8030-cc249e7464b2" containerName="mariadb-account-create" Nov 24 10:22:13 crc kubenswrapper[4944]: E1124 10:22:13.922410 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8269344b-cb43-4154-8550-84147c6deb7d" containerName="mariadb-account-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922420 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8269344b-cb43-4154-8550-84147c6deb7d" containerName="mariadb-account-create" Nov 24 10:22:13 crc kubenswrapper[4944]: E1124 10:22:13.922435 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c" containerName="mariadb-database-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922440 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c" containerName="mariadb-database-create" Nov 24 10:22:13 crc kubenswrapper[4944]: E1124 10:22:13.922451 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf65b97-ae89-462d-9bde-d694c17f8c46" containerName="mariadb-database-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922457 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf65b97-ae89-462d-9bde-d694c17f8c46" containerName="mariadb-database-create" Nov 24 10:22:13 crc kubenswrapper[4944]: E1124 10:22:13.922471 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae14ddf-150b-4d93-9632-b6f3dff92eea" containerName="mariadb-database-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922477 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae14ddf-150b-4d93-9632-b6f3dff92eea" containerName="mariadb-database-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922653 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c" containerName="mariadb-database-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922665 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae14ddf-150b-4d93-9632-b6f3dff92eea" containerName="mariadb-database-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922677 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8269344b-cb43-4154-8550-84147c6deb7d" containerName="mariadb-account-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922690 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bf65b97-ae89-462d-9bde-d694c17f8c46" containerName="mariadb-database-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922697 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="4490175c-09f1-4ba1-8842-5a9a576ffce1" containerName="mariadb-account-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.922707 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="24c9aee6-8750-4862-8030-cc249e7464b2" containerName="mariadb-account-create" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.923260 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.924997 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.925379 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-ntg4v" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.925939 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 10:22:13 crc kubenswrapper[4944]: I1124 10:22:13.937488 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l44vr"] Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.018787 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pkll\" (UniqueName: \"kubernetes.io/projected/bb15418b-0840-4bdb-92c0-3a8076f9fa71-kube-api-access-4pkll\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.018836 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-config-data\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.019009 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-scripts\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.019279 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.120842 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.120926 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pkll\" (UniqueName: \"kubernetes.io/projected/bb15418b-0840-4bdb-92c0-3a8076f9fa71-kube-api-access-4pkll\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.120953 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-config-data\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.120996 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-scripts\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.125375 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-scripts\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.125648 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-config-data\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.125645 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.148799 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pkll\" (UniqueName: \"kubernetes.io/projected/bb15418b-0840-4bdb-92c0-3a8076f9fa71-kube-api-access-4pkll\") pod \"nova-cell0-conductor-db-sync-l44vr\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.250711 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:14 crc kubenswrapper[4944]: I1124 10:22:14.685591 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l44vr"] Nov 24 10:22:14 crc kubenswrapper[4944]: W1124 10:22:14.687439 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb15418b_0840_4bdb_92c0_3a8076f9fa71.slice/crio-50a537b0773a031523d5e4b1d7da9e39b5f3d8649649823e82243f3a342b41ed WatchSource:0}: Error finding container 50a537b0773a031523d5e4b1d7da9e39b5f3d8649649823e82243f3a342b41ed: Status 404 returned error can't find the container with id 50a537b0773a031523d5e4b1d7da9e39b5f3d8649649823e82243f3a342b41ed Nov 24 10:22:15 crc kubenswrapper[4944]: I1124 10:22:15.533326 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l44vr" event={"ID":"bb15418b-0840-4bdb-92c0-3a8076f9fa71","Type":"ContainerStarted","Data":"82e61cc90223698151e9f7f0d34f7b6f6d39f25a0972101eb1dc88b883f153e8"} Nov 24 10:22:15 crc kubenswrapper[4944]: I1124 10:22:15.535522 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l44vr" event={"ID":"bb15418b-0840-4bdb-92c0-3a8076f9fa71","Type":"ContainerStarted","Data":"50a537b0773a031523d5e4b1d7da9e39b5f3d8649649823e82243f3a342b41ed"} Nov 24 10:22:15 crc kubenswrapper[4944]: I1124 10:22:15.569737 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-l44vr" podStartSLOduration=2.569709259 podStartE2EDuration="2.569709259s" podCreationTimestamp="2025-11-24 10:22:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:15.561118375 +0000 UTC m=+5396.095558847" watchObservedRunningTime="2025-11-24 10:22:15.569709259 +0000 UTC m=+5396.104149721" Nov 24 10:22:20 crc kubenswrapper[4944]: I1124 10:22:20.581826 4944 generic.go:334] "Generic (PLEG): container finished" podID="bb15418b-0840-4bdb-92c0-3a8076f9fa71" containerID="82e61cc90223698151e9f7f0d34f7b6f6d39f25a0972101eb1dc88b883f153e8" exitCode=0 Nov 24 10:22:20 crc kubenswrapper[4944]: I1124 10:22:20.581909 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l44vr" event={"ID":"bb15418b-0840-4bdb-92c0-3a8076f9fa71","Type":"ContainerDied","Data":"82e61cc90223698151e9f7f0d34f7b6f6d39f25a0972101eb1dc88b883f153e8"} Nov 24 10:22:21 crc kubenswrapper[4944]: I1124 10:22:21.864393 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:21 crc kubenswrapper[4944]: I1124 10:22:21.953468 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-config-data\") pod \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " Nov 24 10:22:21 crc kubenswrapper[4944]: I1124 10:22:21.953902 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-scripts\") pod \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " Nov 24 10:22:21 crc kubenswrapper[4944]: I1124 10:22:21.953995 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pkll\" (UniqueName: \"kubernetes.io/projected/bb15418b-0840-4bdb-92c0-3a8076f9fa71-kube-api-access-4pkll\") pod \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " Nov 24 10:22:21 crc kubenswrapper[4944]: I1124 10:22:21.954088 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-combined-ca-bundle\") pod \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\" (UID: \"bb15418b-0840-4bdb-92c0-3a8076f9fa71\") " Nov 24 10:22:21 crc kubenswrapper[4944]: I1124 10:22:21.959004 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-scripts" (OuterVolumeSpecName: "scripts") pod "bb15418b-0840-4bdb-92c0-3a8076f9fa71" (UID: "bb15418b-0840-4bdb-92c0-3a8076f9fa71"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:21 crc kubenswrapper[4944]: I1124 10:22:21.968877 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb15418b-0840-4bdb-92c0-3a8076f9fa71-kube-api-access-4pkll" (OuterVolumeSpecName: "kube-api-access-4pkll") pod "bb15418b-0840-4bdb-92c0-3a8076f9fa71" (UID: "bb15418b-0840-4bdb-92c0-3a8076f9fa71"). InnerVolumeSpecName "kube-api-access-4pkll". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:21 crc kubenswrapper[4944]: I1124 10:22:21.980734 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-config-data" (OuterVolumeSpecName: "config-data") pod "bb15418b-0840-4bdb-92c0-3a8076f9fa71" (UID: "bb15418b-0840-4bdb-92c0-3a8076f9fa71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:21 crc kubenswrapper[4944]: I1124 10:22:21.980942 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb15418b-0840-4bdb-92c0-3a8076f9fa71" (UID: "bb15418b-0840-4bdb-92c0-3a8076f9fa71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.056502 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pkll\" (UniqueName: \"kubernetes.io/projected/bb15418b-0840-4bdb-92c0-3a8076f9fa71-kube-api-access-4pkll\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.056533 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.056541 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.056549 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb15418b-0840-4bdb-92c0-3a8076f9fa71-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.599303 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l44vr" event={"ID":"bb15418b-0840-4bdb-92c0-3a8076f9fa71","Type":"ContainerDied","Data":"50a537b0773a031523d5e4b1d7da9e39b5f3d8649649823e82243f3a342b41ed"} Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.599350 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50a537b0773a031523d5e4b1d7da9e39b5f3d8649649823e82243f3a342b41ed" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.599379 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l44vr" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.668906 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 10:22:22 crc kubenswrapper[4944]: E1124 10:22:22.669509 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb15418b-0840-4bdb-92c0-3a8076f9fa71" containerName="nova-cell0-conductor-db-sync" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.669581 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb15418b-0840-4bdb-92c0-3a8076f9fa71" containerName="nova-cell0-conductor-db-sync" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.669872 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb15418b-0840-4bdb-92c0-3a8076f9fa71" containerName="nova-cell0-conductor-db-sync" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.670568 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.672513 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-ntg4v" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.682677 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.684153 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.767553 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.767868 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.768088 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzqrl\" (UniqueName: \"kubernetes.io/projected/f696e4c8-bb0c-410e-af8d-351a42bee463-kube-api-access-hzqrl\") pod \"nova-cell0-conductor-0\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.869336 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzqrl\" (UniqueName: \"kubernetes.io/projected/f696e4c8-bb0c-410e-af8d-351a42bee463-kube-api-access-hzqrl\") pod \"nova-cell0-conductor-0\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.871644 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.871757 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.878025 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.879568 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.885129 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzqrl\" (UniqueName: \"kubernetes.io/projected/f696e4c8-bb0c-410e-af8d-351a42bee463-kube-api-access-hzqrl\") pod \"nova-cell0-conductor-0\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:22 crc kubenswrapper[4944]: I1124 10:22:22.990955 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:23 crc kubenswrapper[4944]: I1124 10:22:23.401064 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 10:22:23 crc kubenswrapper[4944]: I1124 10:22:23.610942 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f696e4c8-bb0c-410e-af8d-351a42bee463","Type":"ContainerStarted","Data":"cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718"} Nov 24 10:22:23 crc kubenswrapper[4944]: I1124 10:22:23.611457 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:23 crc kubenswrapper[4944]: I1124 10:22:23.611471 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f696e4c8-bb0c-410e-af8d-351a42bee463","Type":"ContainerStarted","Data":"0a293a3b644827d3823455ce0def1b3207837c25a0ff7c7b9b4bd8007dbfc7a2"} Nov 24 10:22:23 crc kubenswrapper[4944]: I1124 10:22:23.632992 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.632972517 podStartE2EDuration="1.632972517s" podCreationTimestamp="2025-11-24 10:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:23.627590365 +0000 UTC m=+5404.162030857" watchObservedRunningTime="2025-11-24 10:22:23.632972517 +0000 UTC m=+5404.167412989" Nov 24 10:22:30 crc kubenswrapper[4944]: I1124 10:22:30.836103 4944 scope.go:117] "RemoveContainer" containerID="a4b386881cb0e5796bea4dfee96632381ecf94287bbc539f1958a645e9737fb5" Nov 24 10:22:30 crc kubenswrapper[4944]: I1124 10:22:30.859804 4944 scope.go:117] "RemoveContainer" containerID="e835c2e61896806dff8001fa9a9ff6005548fda86b44d18d38904afa1e7e5107" Nov 24 10:22:30 crc kubenswrapper[4944]: I1124 10:22:30.879347 4944 scope.go:117] "RemoveContainer" containerID="3b7f0f932758a488d457bbd0ecc35528d9eb5c141d69ecce07ad18d0eaf5bfed" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.106619 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.527001 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-bb2gt"] Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.528036 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.531450 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.536629 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.541587 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bb2gt"] Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.547951 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-scripts\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.548011 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n454\" (UniqueName: \"kubernetes.io/projected/f1bbbe09-dbf8-434a-8615-849e02e7873a-kube-api-access-2n454\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.548090 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-config-data\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.548206 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.649762 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n454\" (UniqueName: \"kubernetes.io/projected/f1bbbe09-dbf8-434a-8615-849e02e7873a-kube-api-access-2n454\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.649835 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-config-data\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.649899 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.649992 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-scripts\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.653290 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.654849 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.658254 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-scripts\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.658818 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.668851 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-config-data\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.673403 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.689021 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.700222 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n454\" (UniqueName: \"kubernetes.io/projected/f1bbbe09-dbf8-434a-8615-849e02e7873a-kube-api-access-2n454\") pod \"nova-cell0-cell-mapping-bb2gt\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.751851 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c04de488-e975-4fd3-9adc-fa2d9450d0e2-logs\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.752150 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.752787 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-config-data\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.752939 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnmj7\" (UniqueName: \"kubernetes.io/projected/c04de488-e975-4fd3-9adc-fa2d9450d0e2-kube-api-access-rnmj7\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.789647 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.791509 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.793722 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.809201 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.832010 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.833785 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.837669 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.845764 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.857626 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.858100 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6jxk\" (UniqueName: \"kubernetes.io/projected/563432e9-6f44-4c36-9ab3-37162bc8bd95-kube-api-access-q6jxk\") pod \"nova-scheduler-0\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.858160 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-config-data\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.858185 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnmj7\" (UniqueName: \"kubernetes.io/projected/c04de488-e975-4fd3-9adc-fa2d9450d0e2-kube-api-access-rnmj7\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.858206 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-config-data\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.858232 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3045a69e-a776-42bd-88d9-29a17c401d20-logs\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.858246 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-config-data\") pod \"nova-scheduler-0\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.858287 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhgwd\" (UniqueName: \"kubernetes.io/projected/3045a69e-a776-42bd-88d9-29a17c401d20-kube-api-access-fhgwd\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.858307 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c04de488-e975-4fd3-9adc-fa2d9450d0e2-logs\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.858328 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.858354 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.862534 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.869430 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.870411 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-config-data\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.870782 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c04de488-e975-4fd3-9adc-fa2d9450d0e2-logs\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.923757 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnmj7\" (UniqueName: \"kubernetes.io/projected/c04de488-e975-4fd3-9adc-fa2d9450d0e2-kube-api-access-rnmj7\") pod \"nova-api-0\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " pod="openstack/nova-api-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.954734 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b58cb654c-gwclv"] Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.956732 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.960203 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.960269 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6jxk\" (UniqueName: \"kubernetes.io/projected/563432e9-6f44-4c36-9ab3-37162bc8bd95-kube-api-access-q6jxk\") pod \"nova-scheduler-0\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.960329 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-config-data\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.960358 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3045a69e-a776-42bd-88d9-29a17c401d20-logs\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.960372 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-config-data\") pod \"nova-scheduler-0\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.960420 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhgwd\" (UniqueName: \"kubernetes.io/projected/3045a69e-a776-42bd-88d9-29a17c401d20-kube-api-access-fhgwd\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.960468 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.968219 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3045a69e-a776-42bd-88d9-29a17c401d20-logs\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.976754 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.980272 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.980516 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-config-data\") pod \"nova-scheduler-0\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.986129 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-config-data\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.994488 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhgwd\" (UniqueName: \"kubernetes.io/projected/3045a69e-a776-42bd-88d9-29a17c401d20-kube-api-access-fhgwd\") pod \"nova-metadata-0\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " pod="openstack/nova-metadata-0" Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.996112 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b58cb654c-gwclv"] Nov 24 10:22:33 crc kubenswrapper[4944]: I1124 10:22:33.999021 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6jxk\" (UniqueName: \"kubernetes.io/projected/563432e9-6f44-4c36-9ab3-37162bc8bd95-kube-api-access-q6jxk\") pod \"nova-scheduler-0\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.015313 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.016670 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.018743 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.045662 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.062472 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-nb\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.062525 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.062560 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-config\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.062607 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-964gp\" (UniqueName: \"kubernetes.io/projected/a46345a8-68da-4344-84cc-f535d23c70c7-kube-api-access-964gp\") pod \"nova-cell1-novncproxy-0\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.062639 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.062664 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-sb\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.070465 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mlpf\" (UniqueName: \"kubernetes.io/projected/ea6d6441-ea1d-43df-b4dc-6b838f736b05-kube-api-access-6mlpf\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.070638 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-dns-svc\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.107732 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.128625 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.172916 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mlpf\" (UniqueName: \"kubernetes.io/projected/ea6d6441-ea1d-43df-b4dc-6b838f736b05-kube-api-access-6mlpf\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.172996 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-dns-svc\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.173080 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-nb\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.173103 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.173133 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-config\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.173175 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-964gp\" (UniqueName: \"kubernetes.io/projected/a46345a8-68da-4344-84cc-f535d23c70c7-kube-api-access-964gp\") pod \"nova-cell1-novncproxy-0\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.173205 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.173231 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-sb\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.174176 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-sb\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.175184 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-dns-svc\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.175830 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-nb\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.176967 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-config\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.177473 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.193839 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.195792 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.202993 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-964gp\" (UniqueName: \"kubernetes.io/projected/a46345a8-68da-4344-84cc-f535d23c70c7-kube-api-access-964gp\") pod \"nova-cell1-novncproxy-0\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.204190 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mlpf\" (UniqueName: \"kubernetes.io/projected/ea6d6441-ea1d-43df-b4dc-6b838f736b05-kube-api-access-6mlpf\") pod \"dnsmasq-dns-7b58cb654c-gwclv\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.314457 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.357977 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.430520 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bb2gt"] Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.785693 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bb2gt" event={"ID":"f1bbbe09-dbf8-434a-8615-849e02e7873a","Type":"ContainerStarted","Data":"7f7af4c96b1caf9fab23d6faaf16aa6a1ba05a54131a10006b0041982e74724c"} Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.804447 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.894421 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.945171 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.986683 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fm68v"] Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.990562 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.993872 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 10:22:34 crc kubenswrapper[4944]: I1124 10:22:34.994656 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.016331 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fm68v"] Nov 24 10:22:35 crc kubenswrapper[4944]: W1124 10:22:35.029650 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea6d6441_ea1d_43df_b4dc_6b838f736b05.slice/crio-53b5de04b868cf32b9c9f697180533ee5d36e8e518c0c281da23fb092fc5c2ba WatchSource:0}: Error finding container 53b5de04b868cf32b9c9f697180533ee5d36e8e518c0c281da23fb092fc5c2ba: Status 404 returned error can't find the container with id 53b5de04b868cf32b9c9f697180533ee5d36e8e518c0c281da23fb092fc5c2ba Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.034757 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b58cb654c-gwclv"] Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.119208 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.119348 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cxgf\" (UniqueName: \"kubernetes.io/projected/73011835-f092-4f8c-8ec2-d5928bb99c97-kube-api-access-9cxgf\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.119409 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-config-data\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.119450 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-scripts\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.221086 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cxgf\" (UniqueName: \"kubernetes.io/projected/73011835-f092-4f8c-8ec2-d5928bb99c97-kube-api-access-9cxgf\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.221396 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-config-data\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.221422 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-scripts\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.221486 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.242336 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-config-data\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.243103 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cxgf\" (UniqueName: \"kubernetes.io/projected/73011835-f092-4f8c-8ec2-d5928bb99c97-kube-api-access-9cxgf\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.249148 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.254135 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-scripts\") pod \"nova-cell1-conductor-db-sync-fm68v\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.296984 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.345860 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.802420 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3045a69e-a776-42bd-88d9-29a17c401d20","Type":"ContainerStarted","Data":"fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.802736 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3045a69e-a776-42bd-88d9-29a17c401d20","Type":"ContainerStarted","Data":"99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.802752 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3045a69e-a776-42bd-88d9-29a17c401d20","Type":"ContainerStarted","Data":"9ca04a88421919083e842b11f13534352ac37330e080d7a4a18566816e2414c0"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.808703 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bb2gt" event={"ID":"f1bbbe09-dbf8-434a-8615-849e02e7873a","Type":"ContainerStarted","Data":"0e722ee180ad0881308285bd522a6caafbf07cdb300f271b5e59c675789ee2e3"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.812817 4944 generic.go:334] "Generic (PLEG): container finished" podID="ea6d6441-ea1d-43df-b4dc-6b838f736b05" containerID="5eff8c1c55bd1a723ba72052f7514edbe5625d7d219e61e1046cbc0f56bfc299" exitCode=0 Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.813017 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" event={"ID":"ea6d6441-ea1d-43df-b4dc-6b838f736b05","Type":"ContainerDied","Data":"5eff8c1c55bd1a723ba72052f7514edbe5625d7d219e61e1046cbc0f56bfc299"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.813055 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" event={"ID":"ea6d6441-ea1d-43df-b4dc-6b838f736b05","Type":"ContainerStarted","Data":"53b5de04b868cf32b9c9f697180533ee5d36e8e518c0c281da23fb092fc5c2ba"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.823355 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"563432e9-6f44-4c36-9ab3-37162bc8bd95","Type":"ContainerStarted","Data":"51d3bfffa7c4ee025e73e5b2f38191a9d322849229c0215ab64047ea6c7a0cd6"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.823399 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"563432e9-6f44-4c36-9ab3-37162bc8bd95","Type":"ContainerStarted","Data":"6e5b7e06771c2b3dc4f003f071895e33e100215fa591f19ab04f0a64cdad57bb"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.832109 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a46345a8-68da-4344-84cc-f535d23c70c7","Type":"ContainerStarted","Data":"328b844e22c2e9064ca537cdd1059002c631f209faf697bfa41a5fd590d0cac4"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.832161 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a46345a8-68da-4344-84cc-f535d23c70c7","Type":"ContainerStarted","Data":"7386522d0c1ac192064c05dbd01d6bbd7dcd54dadd417be615cfdcbfedc0ba3d"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.835816 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c04de488-e975-4fd3-9adc-fa2d9450d0e2","Type":"ContainerStarted","Data":"5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.835856 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c04de488-e975-4fd3-9adc-fa2d9450d0e2","Type":"ContainerStarted","Data":"baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.835866 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c04de488-e975-4fd3-9adc-fa2d9450d0e2","Type":"ContainerStarted","Data":"9995f54bfe6a0e2af79d3ea3f464c81e355874f4746b432ff933bb7b69a45d8d"} Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.849566 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.849546504 podStartE2EDuration="2.849546504s" podCreationTimestamp="2025-11-24 10:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:35.830518909 +0000 UTC m=+5416.364959381" watchObservedRunningTime="2025-11-24 10:22:35.849546504 +0000 UTC m=+5416.383986966" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.851448 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fm68v"] Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.859924 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-bb2gt" podStartSLOduration=2.859907274 podStartE2EDuration="2.859907274s" podCreationTimestamp="2025-11-24 10:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:35.846536548 +0000 UTC m=+5416.380977020" watchObservedRunningTime="2025-11-24 10:22:35.859907274 +0000 UTC m=+5416.394347726" Nov 24 10:22:35 crc kubenswrapper[4944]: W1124 10:22:35.863858 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73011835_f092_4f8c_8ec2_d5928bb99c97.slice/crio-97772880b0ab01303c579aa14ca61682c23b5b5299d21f3d90e0a3f5bf012792 WatchSource:0}: Error finding container 97772880b0ab01303c579aa14ca61682c23b5b5299d21f3d90e0a3f5bf012792: Status 404 returned error can't find the container with id 97772880b0ab01303c579aa14ca61682c23b5b5299d21f3d90e0a3f5bf012792 Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.918007 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.917983774 podStartE2EDuration="2.917983774s" podCreationTimestamp="2025-11-24 10:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:35.893816324 +0000 UTC m=+5416.428256796" watchObservedRunningTime="2025-11-24 10:22:35.917983774 +0000 UTC m=+5416.452424236" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.933164 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9331398159999997 podStartE2EDuration="2.933139816s" podCreationTimestamp="2025-11-24 10:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:35.915235967 +0000 UTC m=+5416.449676429" watchObservedRunningTime="2025-11-24 10:22:35.933139816 +0000 UTC m=+5416.467580278" Nov 24 10:22:35 crc kubenswrapper[4944]: I1124 10:22:35.945642 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.945627254 podStartE2EDuration="2.945627254s" podCreationTimestamp="2025-11-24 10:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:35.93794395 +0000 UTC m=+5416.472384412" watchObservedRunningTime="2025-11-24 10:22:35.945627254 +0000 UTC m=+5416.480067716" Nov 24 10:22:36 crc kubenswrapper[4944]: I1124 10:22:36.846550 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" event={"ID":"ea6d6441-ea1d-43df-b4dc-6b838f736b05","Type":"ContainerStarted","Data":"70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff"} Nov 24 10:22:36 crc kubenswrapper[4944]: I1124 10:22:36.847312 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:36 crc kubenswrapper[4944]: I1124 10:22:36.863761 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fm68v" event={"ID":"73011835-f092-4f8c-8ec2-d5928bb99c97","Type":"ContainerStarted","Data":"8c25a2f788ade6a6f856dd1cb31fcc0baa65f54d51b1e64b85890232a5d92ee0"} Nov 24 10:22:36 crc kubenswrapper[4944]: I1124 10:22:36.864117 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fm68v" event={"ID":"73011835-f092-4f8c-8ec2-d5928bb99c97","Type":"ContainerStarted","Data":"97772880b0ab01303c579aa14ca61682c23b5b5299d21f3d90e0a3f5bf012792"} Nov 24 10:22:36 crc kubenswrapper[4944]: I1124 10:22:36.884268 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" podStartSLOduration=3.884241792 podStartE2EDuration="3.884241792s" podCreationTimestamp="2025-11-24 10:22:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:36.880299516 +0000 UTC m=+5417.414739988" watchObservedRunningTime="2025-11-24 10:22:36.884241792 +0000 UTC m=+5417.418682254" Nov 24 10:22:36 crc kubenswrapper[4944]: I1124 10:22:36.915495 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-fm68v" podStartSLOduration=2.915470275 podStartE2EDuration="2.915470275s" podCreationTimestamp="2025-11-24 10:22:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:36.89768613 +0000 UTC m=+5417.432126602" watchObservedRunningTime="2025-11-24 10:22:36.915470275 +0000 UTC m=+5417.449910757" Nov 24 10:22:39 crc kubenswrapper[4944]: I1124 10:22:39.131922 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 10:22:39 crc kubenswrapper[4944]: I1124 10:22:39.132358 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 10:22:39 crc kubenswrapper[4944]: I1124 10:22:39.178672 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 10:22:39 crc kubenswrapper[4944]: I1124 10:22:39.359030 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:39 crc kubenswrapper[4944]: I1124 10:22:39.877266 4944 generic.go:334] "Generic (PLEG): container finished" podID="73011835-f092-4f8c-8ec2-d5928bb99c97" containerID="8c25a2f788ade6a6f856dd1cb31fcc0baa65f54d51b1e64b85890232a5d92ee0" exitCode=0 Nov 24 10:22:39 crc kubenswrapper[4944]: I1124 10:22:39.877342 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fm68v" event={"ID":"73011835-f092-4f8c-8ec2-d5928bb99c97","Type":"ContainerDied","Data":"8c25a2f788ade6a6f856dd1cb31fcc0baa65f54d51b1e64b85890232a5d92ee0"} Nov 24 10:22:40 crc kubenswrapper[4944]: I1124 10:22:40.890945 4944 generic.go:334] "Generic (PLEG): container finished" podID="f1bbbe09-dbf8-434a-8615-849e02e7873a" containerID="0e722ee180ad0881308285bd522a6caafbf07cdb300f271b5e59c675789ee2e3" exitCode=0 Nov 24 10:22:40 crc kubenswrapper[4944]: I1124 10:22:40.891885 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bb2gt" event={"ID":"f1bbbe09-dbf8-434a-8615-849e02e7873a","Type":"ContainerDied","Data":"0e722ee180ad0881308285bd522a6caafbf07cdb300f271b5e59c675789ee2e3"} Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.229467 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.371760 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-combined-ca-bundle\") pod \"73011835-f092-4f8c-8ec2-d5928bb99c97\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.371809 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cxgf\" (UniqueName: \"kubernetes.io/projected/73011835-f092-4f8c-8ec2-d5928bb99c97-kube-api-access-9cxgf\") pod \"73011835-f092-4f8c-8ec2-d5928bb99c97\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.371922 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-config-data\") pod \"73011835-f092-4f8c-8ec2-d5928bb99c97\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.372410 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-scripts\") pod \"73011835-f092-4f8c-8ec2-d5928bb99c97\" (UID: \"73011835-f092-4f8c-8ec2-d5928bb99c97\") " Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.380711 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-scripts" (OuterVolumeSpecName: "scripts") pod "73011835-f092-4f8c-8ec2-d5928bb99c97" (UID: "73011835-f092-4f8c-8ec2-d5928bb99c97"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.381723 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73011835-f092-4f8c-8ec2-d5928bb99c97-kube-api-access-9cxgf" (OuterVolumeSpecName: "kube-api-access-9cxgf") pod "73011835-f092-4f8c-8ec2-d5928bb99c97" (UID: "73011835-f092-4f8c-8ec2-d5928bb99c97"). InnerVolumeSpecName "kube-api-access-9cxgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.404713 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73011835-f092-4f8c-8ec2-d5928bb99c97" (UID: "73011835-f092-4f8c-8ec2-d5928bb99c97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.412405 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-config-data" (OuterVolumeSpecName: "config-data") pod "73011835-f092-4f8c-8ec2-d5928bb99c97" (UID: "73011835-f092-4f8c-8ec2-d5928bb99c97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.477212 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.477243 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cxgf\" (UniqueName: \"kubernetes.io/projected/73011835-f092-4f8c-8ec2-d5928bb99c97-kube-api-access-9cxgf\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.477253 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.477261 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73011835-f092-4f8c-8ec2-d5928bb99c97-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.902873 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fm68v" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.902876 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fm68v" event={"ID":"73011835-f092-4f8c-8ec2-d5928bb99c97","Type":"ContainerDied","Data":"97772880b0ab01303c579aa14ca61682c23b5b5299d21f3d90e0a3f5bf012792"} Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.902944 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97772880b0ab01303c579aa14ca61682c23b5b5299d21f3d90e0a3f5bf012792" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.971314 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 10:22:41 crc kubenswrapper[4944]: E1124 10:22:41.971705 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73011835-f092-4f8c-8ec2-d5928bb99c97" containerName="nova-cell1-conductor-db-sync" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.971723 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="73011835-f092-4f8c-8ec2-d5928bb99c97" containerName="nova-cell1-conductor-db-sync" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.971885 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="73011835-f092-4f8c-8ec2-d5928bb99c97" containerName="nova-cell1-conductor-db-sync" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.972548 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.975002 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 10:22:41 crc kubenswrapper[4944]: I1124 10:22:41.981282 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.088415 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.088467 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzv4r\" (UniqueName: \"kubernetes.io/projected/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-kube-api-access-dzv4r\") pod \"nova-cell1-conductor-0\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.088600 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.189662 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.190115 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.190141 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzv4r\" (UniqueName: \"kubernetes.io/projected/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-kube-api-access-dzv4r\") pod \"nova-cell1-conductor-0\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.196933 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.197120 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.206281 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzv4r\" (UniqueName: \"kubernetes.io/projected/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-kube-api-access-dzv4r\") pod \"nova-cell1-conductor-0\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.273297 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.287673 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.394815 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n454\" (UniqueName: \"kubernetes.io/projected/f1bbbe09-dbf8-434a-8615-849e02e7873a-kube-api-access-2n454\") pod \"f1bbbe09-dbf8-434a-8615-849e02e7873a\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.394893 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-config-data\") pod \"f1bbbe09-dbf8-434a-8615-849e02e7873a\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.395069 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-scripts\") pod \"f1bbbe09-dbf8-434a-8615-849e02e7873a\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.395137 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-combined-ca-bundle\") pod \"f1bbbe09-dbf8-434a-8615-849e02e7873a\" (UID: \"f1bbbe09-dbf8-434a-8615-849e02e7873a\") " Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.399350 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-scripts" (OuterVolumeSpecName: "scripts") pod "f1bbbe09-dbf8-434a-8615-849e02e7873a" (UID: "f1bbbe09-dbf8-434a-8615-849e02e7873a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.402465 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1bbbe09-dbf8-434a-8615-849e02e7873a-kube-api-access-2n454" (OuterVolumeSpecName: "kube-api-access-2n454") pod "f1bbbe09-dbf8-434a-8615-849e02e7873a" (UID: "f1bbbe09-dbf8-434a-8615-849e02e7873a"). InnerVolumeSpecName "kube-api-access-2n454". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.429636 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-config-data" (OuterVolumeSpecName: "config-data") pod "f1bbbe09-dbf8-434a-8615-849e02e7873a" (UID: "f1bbbe09-dbf8-434a-8615-849e02e7873a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.439679 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1bbbe09-dbf8-434a-8615-849e02e7873a" (UID: "f1bbbe09-dbf8-434a-8615-849e02e7873a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.496913 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n454\" (UniqueName: \"kubernetes.io/projected/f1bbbe09-dbf8-434a-8615-849e02e7873a-kube-api-access-2n454\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.497276 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.497287 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.497298 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1bbbe09-dbf8-434a-8615-849e02e7873a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.736677 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 10:22:42 crc kubenswrapper[4944]: W1124 10:22:42.746254 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ddeb86e_6b6a_436f_929c_f53dca84ff3a.slice/crio-5446698ae40f45dc173acca166e8dfe4c979116e9f277f406461aa336975cce6 WatchSource:0}: Error finding container 5446698ae40f45dc173acca166e8dfe4c979116e9f277f406461aa336975cce6: Status 404 returned error can't find the container with id 5446698ae40f45dc173acca166e8dfe4c979116e9f277f406461aa336975cce6 Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.913138 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5ddeb86e-6b6a-436f-929c-f53dca84ff3a","Type":"ContainerStarted","Data":"998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21"} Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.913504 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.913523 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5ddeb86e-6b6a-436f-929c-f53dca84ff3a","Type":"ContainerStarted","Data":"5446698ae40f45dc173acca166e8dfe4c979116e9f277f406461aa336975cce6"} Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.914901 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bb2gt" event={"ID":"f1bbbe09-dbf8-434a-8615-849e02e7873a","Type":"ContainerDied","Data":"7f7af4c96b1caf9fab23d6faaf16aa6a1ba05a54131a10006b0041982e74724c"} Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.914966 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f7af4c96b1caf9fab23d6faaf16aa6a1ba05a54131a10006b0041982e74724c" Nov 24 10:22:42 crc kubenswrapper[4944]: I1124 10:22:42.914930 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bb2gt" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:42.945672 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.9456546970000002 podStartE2EDuration="1.945654697s" podCreationTimestamp="2025-11-24 10:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:42.92564551 +0000 UTC m=+5423.460085972" watchObservedRunningTime="2025-11-24 10:22:42.945654697 +0000 UTC m=+5423.480095159" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.123013 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.123851 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c04de488-e975-4fd3-9adc-fa2d9450d0e2" containerName="nova-api-log" containerID="cri-o://baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4" gracePeriod=30 Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.124297 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c04de488-e975-4fd3-9adc-fa2d9450d0e2" containerName="nova-api-api" containerID="cri-o://5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140" gracePeriod=30 Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.137266 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.137517 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3045a69e-a776-42bd-88d9-29a17c401d20" containerName="nova-metadata-log" containerID="cri-o://99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c" gracePeriod=30 Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.137730 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3045a69e-a776-42bd-88d9-29a17c401d20" containerName="nova-metadata-metadata" containerID="cri-o://fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f" gracePeriod=30 Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.147840 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.148094 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="563432e9-6f44-4c36-9ab3-37162bc8bd95" containerName="nova-scheduler-scheduler" containerID="cri-o://51d3bfffa7c4ee025e73e5b2f38191a9d322849229c0215ab64047ea6c7a0cd6" gracePeriod=30 Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.732765 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.741500 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.924447 4944 generic.go:334] "Generic (PLEG): container finished" podID="c04de488-e975-4fd3-9adc-fa2d9450d0e2" containerID="5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140" exitCode=0 Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.925449 4944 generic.go:334] "Generic (PLEG): container finished" podID="c04de488-e975-4fd3-9adc-fa2d9450d0e2" containerID="baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4" exitCode=143 Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.924536 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c04de488-e975-4fd3-9adc-fa2d9450d0e2","Type":"ContainerDied","Data":"5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140"} Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.925691 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3045a69e-a776-42bd-88d9-29a17c401d20-logs" (OuterVolumeSpecName: "logs") pod "3045a69e-a776-42bd-88d9-29a17c401d20" (UID: "3045a69e-a776-42bd-88d9-29a17c401d20"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.924513 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.925705 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c04de488-e975-4fd3-9adc-fa2d9450d0e2","Type":"ContainerDied","Data":"baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4"} Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.925857 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c04de488-e975-4fd3-9adc-fa2d9450d0e2","Type":"ContainerDied","Data":"9995f54bfe6a0e2af79d3ea3f464c81e355874f4746b432ff933bb7b69a45d8d"} Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.925206 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3045a69e-a776-42bd-88d9-29a17c401d20-logs\") pod \"3045a69e-a776-42bd-88d9-29a17c401d20\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.925722 4944 scope.go:117] "RemoveContainer" containerID="5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.925953 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-config-data\") pod \"3045a69e-a776-42bd-88d9-29a17c401d20\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.926013 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-config-data\") pod \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.926134 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnmj7\" (UniqueName: \"kubernetes.io/projected/c04de488-e975-4fd3-9adc-fa2d9450d0e2-kube-api-access-rnmj7\") pod \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.926173 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-combined-ca-bundle\") pod \"3045a69e-a776-42bd-88d9-29a17c401d20\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.926234 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhgwd\" (UniqueName: \"kubernetes.io/projected/3045a69e-a776-42bd-88d9-29a17c401d20-kube-api-access-fhgwd\") pod \"3045a69e-a776-42bd-88d9-29a17c401d20\" (UID: \"3045a69e-a776-42bd-88d9-29a17c401d20\") " Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.926306 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-combined-ca-bundle\") pod \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.926348 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c04de488-e975-4fd3-9adc-fa2d9450d0e2-logs\") pod \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\" (UID: \"c04de488-e975-4fd3-9adc-fa2d9450d0e2\") " Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.926982 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3045a69e-a776-42bd-88d9-29a17c401d20-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.927164 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c04de488-e975-4fd3-9adc-fa2d9450d0e2-logs" (OuterVolumeSpecName: "logs") pod "c04de488-e975-4fd3-9adc-fa2d9450d0e2" (UID: "c04de488-e975-4fd3-9adc-fa2d9450d0e2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.931265 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3045a69e-a776-42bd-88d9-29a17c401d20-kube-api-access-fhgwd" (OuterVolumeSpecName: "kube-api-access-fhgwd") pod "3045a69e-a776-42bd-88d9-29a17c401d20" (UID: "3045a69e-a776-42bd-88d9-29a17c401d20"). InnerVolumeSpecName "kube-api-access-fhgwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.932866 4944 generic.go:334] "Generic (PLEG): container finished" podID="3045a69e-a776-42bd-88d9-29a17c401d20" containerID="fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f" exitCode=0 Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.932894 4944 generic.go:334] "Generic (PLEG): container finished" podID="3045a69e-a776-42bd-88d9-29a17c401d20" containerID="99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c" exitCode=143 Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.932913 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3045a69e-a776-42bd-88d9-29a17c401d20","Type":"ContainerDied","Data":"fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f"} Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.932934 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.932964 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3045a69e-a776-42bd-88d9-29a17c401d20","Type":"ContainerDied","Data":"99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c"} Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.932981 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3045a69e-a776-42bd-88d9-29a17c401d20","Type":"ContainerDied","Data":"9ca04a88421919083e842b11f13534352ac37330e080d7a4a18566816e2414c0"} Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.952971 4944 scope.go:117] "RemoveContainer" containerID="baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.953772 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c04de488-e975-4fd3-9adc-fa2d9450d0e2-kube-api-access-rnmj7" (OuterVolumeSpecName: "kube-api-access-rnmj7") pod "c04de488-e975-4fd3-9adc-fa2d9450d0e2" (UID: "c04de488-e975-4fd3-9adc-fa2d9450d0e2"). InnerVolumeSpecName "kube-api-access-rnmj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.964104 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3045a69e-a776-42bd-88d9-29a17c401d20" (UID: "3045a69e-a776-42bd-88d9-29a17c401d20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.964118 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c04de488-e975-4fd3-9adc-fa2d9450d0e2" (UID: "c04de488-e975-4fd3-9adc-fa2d9450d0e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.966877 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-config-data" (OuterVolumeSpecName: "config-data") pod "3045a69e-a776-42bd-88d9-29a17c401d20" (UID: "3045a69e-a776-42bd-88d9-29a17c401d20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.969275 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-config-data" (OuterVolumeSpecName: "config-data") pod "c04de488-e975-4fd3-9adc-fa2d9450d0e2" (UID: "c04de488-e975-4fd3-9adc-fa2d9450d0e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.975326 4944 scope.go:117] "RemoveContainer" containerID="5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140" Nov 24 10:22:43 crc kubenswrapper[4944]: E1124 10:22:43.975852 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140\": container with ID starting with 5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140 not found: ID does not exist" containerID="5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.975904 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140"} err="failed to get container status \"5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140\": rpc error: code = NotFound desc = could not find container \"5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140\": container with ID starting with 5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140 not found: ID does not exist" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.975931 4944 scope.go:117] "RemoveContainer" containerID="baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4" Nov 24 10:22:43 crc kubenswrapper[4944]: E1124 10:22:43.976575 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4\": container with ID starting with baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4 not found: ID does not exist" containerID="baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.976668 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4"} err="failed to get container status \"baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4\": rpc error: code = NotFound desc = could not find container \"baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4\": container with ID starting with baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4 not found: ID does not exist" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.976697 4944 scope.go:117] "RemoveContainer" containerID="5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.977025 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140"} err="failed to get container status \"5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140\": rpc error: code = NotFound desc = could not find container \"5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140\": container with ID starting with 5daf263f245b60702ea56cc71e7d0c714332c49ca3802710a2aba90264aa8140 not found: ID does not exist" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.977058 4944 scope.go:117] "RemoveContainer" containerID="baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.977380 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4"} err="failed to get container status \"baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4\": rpc error: code = NotFound desc = could not find container \"baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4\": container with ID starting with baacf63a626ef7ac7c3a2388f7bda1504d149bf973d20711ac82dec842a759b4 not found: ID does not exist" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.977400 4944 scope.go:117] "RemoveContainer" containerID="fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f" Nov 24 10:22:43 crc kubenswrapper[4944]: I1124 10:22:43.995118 4944 scope.go:117] "RemoveContainer" containerID="99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.010388 4944 scope.go:117] "RemoveContainer" containerID="fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f" Nov 24 10:22:44 crc kubenswrapper[4944]: E1124 10:22:44.010853 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f\": container with ID starting with fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f not found: ID does not exist" containerID="fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.010911 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f"} err="failed to get container status \"fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f\": rpc error: code = NotFound desc = could not find container \"fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f\": container with ID starting with fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f not found: ID does not exist" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.010938 4944 scope.go:117] "RemoveContainer" containerID="99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c" Nov 24 10:22:44 crc kubenswrapper[4944]: E1124 10:22:44.011279 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c\": container with ID starting with 99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c not found: ID does not exist" containerID="99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.011310 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c"} err="failed to get container status \"99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c\": rpc error: code = NotFound desc = could not find container \"99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c\": container with ID starting with 99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c not found: ID does not exist" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.011336 4944 scope.go:117] "RemoveContainer" containerID="fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.011634 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f"} err="failed to get container status \"fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f\": rpc error: code = NotFound desc = could not find container \"fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f\": container with ID starting with fe72a56edd9669a9598d7f71796028bd6c642120683e8cc96553eb43c994c72f not found: ID does not exist" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.011656 4944 scope.go:117] "RemoveContainer" containerID="99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.011835 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c"} err="failed to get container status \"99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c\": rpc error: code = NotFound desc = could not find container \"99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c\": container with ID starting with 99c654975d1f91d89db20e98ed111e2efac06fa82438a683497da688b5fe609c not found: ID does not exist" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.028004 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.028024 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.028034 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnmj7\" (UniqueName: \"kubernetes.io/projected/c04de488-e975-4fd3-9adc-fa2d9450d0e2-kube-api-access-rnmj7\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.028056 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3045a69e-a776-42bd-88d9-29a17c401d20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.028067 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhgwd\" (UniqueName: \"kubernetes.io/projected/3045a69e-a776-42bd-88d9-29a17c401d20-kube-api-access-fhgwd\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.028075 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c04de488-e975-4fd3-9adc-fa2d9450d0e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.028083 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c04de488-e975-4fd3-9adc-fa2d9450d0e2-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.302470 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.307933 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.314622 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.316745 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.320798 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.327277 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 10:22:44 crc kubenswrapper[4944]: E1124 10:22:44.327645 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3045a69e-a776-42bd-88d9-29a17c401d20" containerName="nova-metadata-metadata" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.327661 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3045a69e-a776-42bd-88d9-29a17c401d20" containerName="nova-metadata-metadata" Nov 24 10:22:44 crc kubenswrapper[4944]: E1124 10:22:44.327670 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1bbbe09-dbf8-434a-8615-849e02e7873a" containerName="nova-manage" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.327676 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1bbbe09-dbf8-434a-8615-849e02e7873a" containerName="nova-manage" Nov 24 10:22:44 crc kubenswrapper[4944]: E1124 10:22:44.327734 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3045a69e-a776-42bd-88d9-29a17c401d20" containerName="nova-metadata-log" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.327740 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3045a69e-a776-42bd-88d9-29a17c401d20" containerName="nova-metadata-log" Nov 24 10:22:44 crc kubenswrapper[4944]: E1124 10:22:44.327754 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c04de488-e975-4fd3-9adc-fa2d9450d0e2" containerName="nova-api-api" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.327760 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c04de488-e975-4fd3-9adc-fa2d9450d0e2" containerName="nova-api-api" Nov 24 10:22:44 crc kubenswrapper[4944]: E1124 10:22:44.327769 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c04de488-e975-4fd3-9adc-fa2d9450d0e2" containerName="nova-api-log" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.327775 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c04de488-e975-4fd3-9adc-fa2d9450d0e2" containerName="nova-api-log" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.327941 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="c04de488-e975-4fd3-9adc-fa2d9450d0e2" containerName="nova-api-log" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.327960 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3045a69e-a776-42bd-88d9-29a17c401d20" containerName="nova-metadata-metadata" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.327974 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="c04de488-e975-4fd3-9adc-fa2d9450d0e2" containerName="nova-api-api" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.327981 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3045a69e-a776-42bd-88d9-29a17c401d20" containerName="nova-metadata-log" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.327993 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1bbbe09-dbf8-434a-8615-849e02e7873a" containerName="nova-manage" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.329082 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.334775 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.355975 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.358994 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.400946 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.417058 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.418719 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.421198 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.433999 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-config-data\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.434064 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t77hc\" (UniqueName: \"kubernetes.io/projected/12e5c569-ab01-4247-8b57-7c3616ac81af-kube-api-access-t77hc\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.434154 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.434197 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12e5c569-ab01-4247-8b57-7c3616ac81af-logs\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.449721 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.485685 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f7979997-bz56v"] Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.485936 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f7979997-bz56v" podUID="7b566b74-289c-4e74-9708-01730c66c939" containerName="dnsmasq-dns" containerID="cri-o://70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf" gracePeriod=10 Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.535583 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12e5c569-ab01-4247-8b57-7c3616ac81af-logs\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.535839 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/984463ec-0877-4461-a2a3-7d677c4ea1bb-logs\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.535926 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-config-data\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.535961 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t77hc\" (UniqueName: \"kubernetes.io/projected/12e5c569-ab01-4247-8b57-7c3616ac81af-kube-api-access-t77hc\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.535992 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.536031 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-config-data\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.536073 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzpnw\" (UniqueName: \"kubernetes.io/projected/984463ec-0877-4461-a2a3-7d677c4ea1bb-kube-api-access-pzpnw\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.536086 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12e5c569-ab01-4247-8b57-7c3616ac81af-logs\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.536109 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.542918 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-config-data\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.545575 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.565760 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t77hc\" (UniqueName: \"kubernetes.io/projected/12e5c569-ab01-4247-8b57-7c3616ac81af-kube-api-access-t77hc\") pod \"nova-api-0\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.637977 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.638302 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-config-data\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.638365 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzpnw\" (UniqueName: \"kubernetes.io/projected/984463ec-0877-4461-a2a3-7d677c4ea1bb-kube-api-access-pzpnw\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.638551 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/984463ec-0877-4461-a2a3-7d677c4ea1bb-logs\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.639065 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/984463ec-0877-4461-a2a3-7d677c4ea1bb-logs\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.643493 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-config-data\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.649228 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.657714 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzpnw\" (UniqueName: \"kubernetes.io/projected/984463ec-0877-4461-a2a3-7d677c4ea1bb-kube-api-access-pzpnw\") pod \"nova-metadata-0\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.700937 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.745016 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.938463 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.942916 4944 generic.go:334] "Generic (PLEG): container finished" podID="7b566b74-289c-4e74-9708-01730c66c939" containerID="70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf" exitCode=0 Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.942966 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7979997-bz56v" event={"ID":"7b566b74-289c-4e74-9708-01730c66c939","Type":"ContainerDied","Data":"70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf"} Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.942989 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7979997-bz56v" event={"ID":"7b566b74-289c-4e74-9708-01730c66c939","Type":"ContainerDied","Data":"b47cfba1043ca125d51c9298923f72f3942012e0463ff03cf70138ea11bc5aea"} Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.943007 4944 scope.go:117] "RemoveContainer" containerID="70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.943124 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7979997-bz56v" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.967811 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:22:44 crc kubenswrapper[4944]: I1124 10:22:44.978751 4944 scope.go:117] "RemoveContainer" containerID="4fe194bc8caa34e0fbff1faaea361298d8c70c845d53bf961c5512922c724953" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.016578 4944 scope.go:117] "RemoveContainer" containerID="70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf" Nov 24 10:22:45 crc kubenswrapper[4944]: E1124 10:22:45.017464 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf\": container with ID starting with 70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf not found: ID does not exist" containerID="70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.017504 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf"} err="failed to get container status \"70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf\": rpc error: code = NotFound desc = could not find container \"70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf\": container with ID starting with 70cd51b9c419c587036650f46279113b64f0fee269c9f935a90a4406fffbb8bf not found: ID does not exist" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.017534 4944 scope.go:117] "RemoveContainer" containerID="4fe194bc8caa34e0fbff1faaea361298d8c70c845d53bf961c5512922c724953" Nov 24 10:22:45 crc kubenswrapper[4944]: E1124 10:22:45.017820 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fe194bc8caa34e0fbff1faaea361298d8c70c845d53bf961c5512922c724953\": container with ID starting with 4fe194bc8caa34e0fbff1faaea361298d8c70c845d53bf961c5512922c724953 not found: ID does not exist" containerID="4fe194bc8caa34e0fbff1faaea361298d8c70c845d53bf961c5512922c724953" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.017865 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe194bc8caa34e0fbff1faaea361298d8c70c845d53bf961c5512922c724953"} err="failed to get container status \"4fe194bc8caa34e0fbff1faaea361298d8c70c845d53bf961c5512922c724953\": rpc error: code = NotFound desc = could not find container \"4fe194bc8caa34e0fbff1faaea361298d8c70c845d53bf961c5512922c724953\": container with ID starting with 4fe194bc8caa34e0fbff1faaea361298d8c70c845d53bf961c5512922c724953 not found: ID does not exist" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.046071 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-sb\") pod \"7b566b74-289c-4e74-9708-01730c66c939\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.046156 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzp9c\" (UniqueName: \"kubernetes.io/projected/7b566b74-289c-4e74-9708-01730c66c939-kube-api-access-rzp9c\") pod \"7b566b74-289c-4e74-9708-01730c66c939\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.047799 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-nb\") pod \"7b566b74-289c-4e74-9708-01730c66c939\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.047872 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-config\") pod \"7b566b74-289c-4e74-9708-01730c66c939\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.047907 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-dns-svc\") pod \"7b566b74-289c-4e74-9708-01730c66c939\" (UID: \"7b566b74-289c-4e74-9708-01730c66c939\") " Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.063089 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b566b74-289c-4e74-9708-01730c66c939-kube-api-access-rzp9c" (OuterVolumeSpecName: "kube-api-access-rzp9c") pod "7b566b74-289c-4e74-9708-01730c66c939" (UID: "7b566b74-289c-4e74-9708-01730c66c939"). InnerVolumeSpecName "kube-api-access-rzp9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.105744 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7b566b74-289c-4e74-9708-01730c66c939" (UID: "7b566b74-289c-4e74-9708-01730c66c939"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.117201 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7b566b74-289c-4e74-9708-01730c66c939" (UID: "7b566b74-289c-4e74-9708-01730c66c939"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.120166 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-config" (OuterVolumeSpecName: "config") pod "7b566b74-289c-4e74-9708-01730c66c939" (UID: "7b566b74-289c-4e74-9708-01730c66c939"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.135763 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7b566b74-289c-4e74-9708-01730c66c939" (UID: "7b566b74-289c-4e74-9708-01730c66c939"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.150406 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzp9c\" (UniqueName: \"kubernetes.io/projected/7b566b74-289c-4e74-9708-01730c66c939-kube-api-access-rzp9c\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.150440 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.150451 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.150459 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.150467 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b566b74-289c-4e74-9708-01730c66c939-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.233928 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.297967 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.314969 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f7979997-bz56v"] Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.326083 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f7979997-bz56v"] Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.957259 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12e5c569-ab01-4247-8b57-7c3616ac81af","Type":"ContainerStarted","Data":"ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce"} Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.957716 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12e5c569-ab01-4247-8b57-7c3616ac81af","Type":"ContainerStarted","Data":"fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a"} Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.957731 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12e5c569-ab01-4247-8b57-7c3616ac81af","Type":"ContainerStarted","Data":"e5e5eb89b85f963794ef80a032bb4abde0baf2d0306372a626d4202d5f1b265b"} Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.959457 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"984463ec-0877-4461-a2a3-7d677c4ea1bb","Type":"ContainerStarted","Data":"696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd"} Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.959500 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"984463ec-0877-4461-a2a3-7d677c4ea1bb","Type":"ContainerStarted","Data":"4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2"} Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.959514 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"984463ec-0877-4461-a2a3-7d677c4ea1bb","Type":"ContainerStarted","Data":"919ac8a4773614940a41fc1a857a68114f28082725f0d52627802a19f53ef0d5"} Nov 24 10:22:45 crc kubenswrapper[4944]: I1124 10:22:45.982086 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.982062832 podStartE2EDuration="1.982062832s" podCreationTimestamp="2025-11-24 10:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:45.976816725 +0000 UTC m=+5426.511257197" watchObservedRunningTime="2025-11-24 10:22:45.982062832 +0000 UTC m=+5426.516503294" Nov 24 10:22:46 crc kubenswrapper[4944]: I1124 10:22:46.013629 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.013602496 podStartE2EDuration="2.013602496s" podCreationTimestamp="2025-11-24 10:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:46.006354305 +0000 UTC m=+5426.540794777" watchObservedRunningTime="2025-11-24 10:22:46.013602496 +0000 UTC m=+5426.548042968" Nov 24 10:22:46 crc kubenswrapper[4944]: I1124 10:22:46.285804 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3045a69e-a776-42bd-88d9-29a17c401d20" path="/var/lib/kubelet/pods/3045a69e-a776-42bd-88d9-29a17c401d20/volumes" Nov 24 10:22:46 crc kubenswrapper[4944]: I1124 10:22:46.286406 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b566b74-289c-4e74-9708-01730c66c939" path="/var/lib/kubelet/pods/7b566b74-289c-4e74-9708-01730c66c939/volumes" Nov 24 10:22:46 crc kubenswrapper[4944]: I1124 10:22:46.286982 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c04de488-e975-4fd3-9adc-fa2d9450d0e2" path="/var/lib/kubelet/pods/c04de488-e975-4fd3-9adc-fa2d9450d0e2/volumes" Nov 24 10:22:46 crc kubenswrapper[4944]: I1124 10:22:46.970843 4944 generic.go:334] "Generic (PLEG): container finished" podID="563432e9-6f44-4c36-9ab3-37162bc8bd95" containerID="51d3bfffa7c4ee025e73e5b2f38191a9d322849229c0215ab64047ea6c7a0cd6" exitCode=0 Nov 24 10:22:46 crc kubenswrapper[4944]: I1124 10:22:46.970923 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"563432e9-6f44-4c36-9ab3-37162bc8bd95","Type":"ContainerDied","Data":"51d3bfffa7c4ee025e73e5b2f38191a9d322849229c0215ab64047ea6c7a0cd6"} Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.064412 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.187256 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6jxk\" (UniqueName: \"kubernetes.io/projected/563432e9-6f44-4c36-9ab3-37162bc8bd95-kube-api-access-q6jxk\") pod \"563432e9-6f44-4c36-9ab3-37162bc8bd95\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.187297 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-config-data\") pod \"563432e9-6f44-4c36-9ab3-37162bc8bd95\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.187438 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-combined-ca-bundle\") pod \"563432e9-6f44-4c36-9ab3-37162bc8bd95\" (UID: \"563432e9-6f44-4c36-9ab3-37162bc8bd95\") " Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.193332 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/563432e9-6f44-4c36-9ab3-37162bc8bd95-kube-api-access-q6jxk" (OuterVolumeSpecName: "kube-api-access-q6jxk") pod "563432e9-6f44-4c36-9ab3-37162bc8bd95" (UID: "563432e9-6f44-4c36-9ab3-37162bc8bd95"). InnerVolumeSpecName "kube-api-access-q6jxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.212415 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-config-data" (OuterVolumeSpecName: "config-data") pod "563432e9-6f44-4c36-9ab3-37162bc8bd95" (UID: "563432e9-6f44-4c36-9ab3-37162bc8bd95"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.217720 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "563432e9-6f44-4c36-9ab3-37162bc8bd95" (UID: "563432e9-6f44-4c36-9ab3-37162bc8bd95"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.289864 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6jxk\" (UniqueName: \"kubernetes.io/projected/563432e9-6f44-4c36-9ab3-37162bc8bd95-kube-api-access-q6jxk\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.289898 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.289909 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563432e9-6f44-4c36-9ab3-37162bc8bd95-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.315773 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.725259 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-g6srq"] Nov 24 10:22:47 crc kubenswrapper[4944]: E1124 10:22:47.726012 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563432e9-6f44-4c36-9ab3-37162bc8bd95" containerName="nova-scheduler-scheduler" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.726037 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="563432e9-6f44-4c36-9ab3-37162bc8bd95" containerName="nova-scheduler-scheduler" Nov 24 10:22:47 crc kubenswrapper[4944]: E1124 10:22:47.726067 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b566b74-289c-4e74-9708-01730c66c939" containerName="dnsmasq-dns" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.726076 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b566b74-289c-4e74-9708-01730c66c939" containerName="dnsmasq-dns" Nov 24 10:22:47 crc kubenswrapper[4944]: E1124 10:22:47.726085 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b566b74-289c-4e74-9708-01730c66c939" containerName="init" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.726093 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b566b74-289c-4e74-9708-01730c66c939" containerName="init" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.726384 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b566b74-289c-4e74-9708-01730c66c939" containerName="dnsmasq-dns" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.726410 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="563432e9-6f44-4c36-9ab3-37162bc8bd95" containerName="nova-scheduler-scheduler" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.727093 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.728781 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.729697 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.735412 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-g6srq"] Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.898974 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-config-data\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.899018 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.899069 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qf4z\" (UniqueName: \"kubernetes.io/projected/8fa3778d-7506-407d-b5d7-b32b310c36d9-kube-api-access-4qf4z\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.899099 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-scripts\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.980409 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"563432e9-6f44-4c36-9ab3-37162bc8bd95","Type":"ContainerDied","Data":"6e5b7e06771c2b3dc4f003f071895e33e100215fa591f19ab04f0a64cdad57bb"} Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.980456 4944 scope.go:117] "RemoveContainer" containerID="51d3bfffa7c4ee025e73e5b2f38191a9d322849229c0215ab64047ea6c7a0cd6" Nov 24 10:22:47 crc kubenswrapper[4944]: I1124 10:22:47.980487 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.000536 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qf4z\" (UniqueName: \"kubernetes.io/projected/8fa3778d-7506-407d-b5d7-b32b310c36d9-kube-api-access-4qf4z\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.000594 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-scripts\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.000714 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-config-data\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.000735 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.018154 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-config-data\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.020612 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qf4z\" (UniqueName: \"kubernetes.io/projected/8fa3778d-7506-407d-b5d7-b32b310c36d9-kube-api-access-4qf4z\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.023896 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-scripts\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.035560 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-g6srq\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.035630 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.044289 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.046540 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.063131 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.064519 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.067875 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.070615 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.203925 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-config-data\") pod \"nova-scheduler-0\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.203983 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxz9x\" (UniqueName: \"kubernetes.io/projected/65f8103b-a3c8-4f50-ba83-70589c965a50-kube-api-access-bxz9x\") pod \"nova-scheduler-0\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.204093 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.287936 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="563432e9-6f44-4c36-9ab3-37162bc8bd95" path="/var/lib/kubelet/pods/563432e9-6f44-4c36-9ab3-37162bc8bd95/volumes" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.306799 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxz9x\" (UniqueName: \"kubernetes.io/projected/65f8103b-a3c8-4f50-ba83-70589c965a50-kube-api-access-bxz9x\") pod \"nova-scheduler-0\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.306882 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.307016 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-config-data\") pod \"nova-scheduler-0\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.313680 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.324784 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-config-data\") pod \"nova-scheduler-0\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.328768 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxz9x\" (UniqueName: \"kubernetes.io/projected/65f8103b-a3c8-4f50-ba83-70589c965a50-kube-api-access-bxz9x\") pod \"nova-scheduler-0\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.469447 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.573642 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-g6srq"] Nov 24 10:22:48 crc kubenswrapper[4944]: W1124 10:22:48.585338 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fa3778d_7506_407d_b5d7_b32b310c36d9.slice/crio-dd38fc20ebe1ce478060cf0e041887c6f3dc8e45342bed8b2257a09069885acb WatchSource:0}: Error finding container dd38fc20ebe1ce478060cf0e041887c6f3dc8e45342bed8b2257a09069885acb: Status 404 returned error can't find the container with id dd38fc20ebe1ce478060cf0e041887c6f3dc8e45342bed8b2257a09069885acb Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.929900 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.993568 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65f8103b-a3c8-4f50-ba83-70589c965a50","Type":"ContainerStarted","Data":"cd858913aa5b1c97e35d11b3b37cc89ac948be82ae4a04389302bbb27d2410bf"} Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.994809 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-g6srq" event={"ID":"8fa3778d-7506-407d-b5d7-b32b310c36d9","Type":"ContainerStarted","Data":"1f07ae27e5791b89cf80fde0600cc7c6e93b782f0e74fb16ca2aeec496f19b2f"} Nov 24 10:22:48 crc kubenswrapper[4944]: I1124 10:22:48.994833 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-g6srq" event={"ID":"8fa3778d-7506-407d-b5d7-b32b310c36d9","Type":"ContainerStarted","Data":"dd38fc20ebe1ce478060cf0e041887c6f3dc8e45342bed8b2257a09069885acb"} Nov 24 10:22:49 crc kubenswrapper[4944]: I1124 10:22:49.011186 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-g6srq" podStartSLOduration=2.011170384 podStartE2EDuration="2.011170384s" podCreationTimestamp="2025-11-24 10:22:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:49.009420038 +0000 UTC m=+5429.543860490" watchObservedRunningTime="2025-11-24 10:22:49.011170384 +0000 UTC m=+5429.545610846" Nov 24 10:22:49 crc kubenswrapper[4944]: I1124 10:22:49.745945 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 10:22:49 crc kubenswrapper[4944]: I1124 10:22:49.746298 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 10:22:50 crc kubenswrapper[4944]: I1124 10:22:50.003444 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65f8103b-a3c8-4f50-ba83-70589c965a50","Type":"ContainerStarted","Data":"cf46c94b591c2e8588364e01dc9d6626d8756a4890a25b09c8d19571a5579653"} Nov 24 10:22:50 crc kubenswrapper[4944]: I1124 10:22:50.025130 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.02511434 podStartE2EDuration="2.02511434s" podCreationTimestamp="2025-11-24 10:22:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:22:50.019614805 +0000 UTC m=+5430.554055267" watchObservedRunningTime="2025-11-24 10:22:50.02511434 +0000 UTC m=+5430.559554802" Nov 24 10:22:53 crc kubenswrapper[4944]: I1124 10:22:53.469889 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 10:22:53 crc kubenswrapper[4944]: I1124 10:22:53.548809 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:22:53 crc kubenswrapper[4944]: I1124 10:22:53.548869 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:22:54 crc kubenswrapper[4944]: I1124 10:22:54.040560 4944 generic.go:334] "Generic (PLEG): container finished" podID="8fa3778d-7506-407d-b5d7-b32b310c36d9" containerID="1f07ae27e5791b89cf80fde0600cc7c6e93b782f0e74fb16ca2aeec496f19b2f" exitCode=0 Nov 24 10:22:54 crc kubenswrapper[4944]: I1124 10:22:54.040649 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-g6srq" event={"ID":"8fa3778d-7506-407d-b5d7-b32b310c36d9","Type":"ContainerDied","Data":"1f07ae27e5791b89cf80fde0600cc7c6e93b782f0e74fb16ca2aeec496f19b2f"} Nov 24 10:22:54 crc kubenswrapper[4944]: I1124 10:22:54.701712 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 10:22:54 crc kubenswrapper[4944]: I1124 10:22:54.702412 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 10:22:54 crc kubenswrapper[4944]: I1124 10:22:54.746218 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 10:22:54 crc kubenswrapper[4944]: I1124 10:22:54.746269 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.401814 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.532727 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qf4z\" (UniqueName: \"kubernetes.io/projected/8fa3778d-7506-407d-b5d7-b32b310c36d9-kube-api-access-4qf4z\") pod \"8fa3778d-7506-407d-b5d7-b32b310c36d9\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.532788 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-scripts\") pod \"8fa3778d-7506-407d-b5d7-b32b310c36d9\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.532830 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-combined-ca-bundle\") pod \"8fa3778d-7506-407d-b5d7-b32b310c36d9\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.532867 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-config-data\") pod \"8fa3778d-7506-407d-b5d7-b32b310c36d9\" (UID: \"8fa3778d-7506-407d-b5d7-b32b310c36d9\") " Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.544291 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fa3778d-7506-407d-b5d7-b32b310c36d9-kube-api-access-4qf4z" (OuterVolumeSpecName: "kube-api-access-4qf4z") pod "8fa3778d-7506-407d-b5d7-b32b310c36d9" (UID: "8fa3778d-7506-407d-b5d7-b32b310c36d9"). InnerVolumeSpecName "kube-api-access-4qf4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.554145 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-scripts" (OuterVolumeSpecName: "scripts") pod "8fa3778d-7506-407d-b5d7-b32b310c36d9" (UID: "8fa3778d-7506-407d-b5d7-b32b310c36d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.558156 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fa3778d-7506-407d-b5d7-b32b310c36d9" (UID: "8fa3778d-7506-407d-b5d7-b32b310c36d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.560298 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-config-data" (OuterVolumeSpecName: "config-data") pod "8fa3778d-7506-407d-b5d7-b32b310c36d9" (UID: "8fa3778d-7506-407d-b5d7-b32b310c36d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.636610 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qf4z\" (UniqueName: \"kubernetes.io/projected/8fa3778d-7506-407d-b5d7-b32b310c36d9-kube-api-access-4qf4z\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.636672 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.636685 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.636700 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa3778d-7506-407d-b5d7-b32b310c36d9-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.785351 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.67:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.785402 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.67:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.867486 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.68:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:22:55 crc kubenswrapper[4944]: I1124 10:22:55.867887 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.68:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:22:56 crc kubenswrapper[4944]: I1124 10:22:56.058236 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-g6srq" event={"ID":"8fa3778d-7506-407d-b5d7-b32b310c36d9","Type":"ContainerDied","Data":"dd38fc20ebe1ce478060cf0e041887c6f3dc8e45342bed8b2257a09069885acb"} Nov 24 10:22:56 crc kubenswrapper[4944]: I1124 10:22:56.058281 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd38fc20ebe1ce478060cf0e041887c6f3dc8e45342bed8b2257a09069885acb" Nov 24 10:22:56 crc kubenswrapper[4944]: I1124 10:22:56.058319 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-g6srq" Nov 24 10:22:56 crc kubenswrapper[4944]: I1124 10:22:56.238607 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:22:56 crc kubenswrapper[4944]: I1124 10:22:56.238815 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerName="nova-api-log" containerID="cri-o://fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a" gracePeriod=30 Nov 24 10:22:56 crc kubenswrapper[4944]: I1124 10:22:56.239275 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerName="nova-api-api" containerID="cri-o://ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce" gracePeriod=30 Nov 24 10:22:56 crc kubenswrapper[4944]: I1124 10:22:56.255722 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:22:56 crc kubenswrapper[4944]: I1124 10:22:56.255954 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="65f8103b-a3c8-4f50-ba83-70589c965a50" containerName="nova-scheduler-scheduler" containerID="cri-o://cf46c94b591c2e8588364e01dc9d6626d8756a4890a25b09c8d19571a5579653" gracePeriod=30 Nov 24 10:22:56 crc kubenswrapper[4944]: I1124 10:22:56.288429 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:22:56 crc kubenswrapper[4944]: I1124 10:22:56.288607 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerName="nova-metadata-log" containerID="cri-o://4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2" gracePeriod=30 Nov 24 10:22:56 crc kubenswrapper[4944]: I1124 10:22:56.288979 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerName="nova-metadata-metadata" containerID="cri-o://696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd" gracePeriod=30 Nov 24 10:22:57 crc kubenswrapper[4944]: I1124 10:22:57.067850 4944 generic.go:334] "Generic (PLEG): container finished" podID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerID="4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2" exitCode=143 Nov 24 10:22:57 crc kubenswrapper[4944]: I1124 10:22:57.067919 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"984463ec-0877-4461-a2a3-7d677c4ea1bb","Type":"ContainerDied","Data":"4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2"} Nov 24 10:22:57 crc kubenswrapper[4944]: I1124 10:22:57.070662 4944 generic.go:334] "Generic (PLEG): container finished" podID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerID="fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a" exitCode=143 Nov 24 10:22:57 crc kubenswrapper[4944]: I1124 10:22:57.070699 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12e5c569-ab01-4247-8b57-7c3616ac81af","Type":"ContainerDied","Data":"fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a"} Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.097952 4944 generic.go:334] "Generic (PLEG): container finished" podID="65f8103b-a3c8-4f50-ba83-70589c965a50" containerID="cf46c94b591c2e8588364e01dc9d6626d8756a4890a25b09c8d19571a5579653" exitCode=0 Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.098008 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65f8103b-a3c8-4f50-ba83-70589c965a50","Type":"ContainerDied","Data":"cf46c94b591c2e8588364e01dc9d6626d8756a4890a25b09c8d19571a5579653"} Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.439821 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.525749 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxz9x\" (UniqueName: \"kubernetes.io/projected/65f8103b-a3c8-4f50-ba83-70589c965a50-kube-api-access-bxz9x\") pod \"65f8103b-a3c8-4f50-ba83-70589c965a50\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.525905 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-config-data\") pod \"65f8103b-a3c8-4f50-ba83-70589c965a50\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.526026 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-combined-ca-bundle\") pod \"65f8103b-a3c8-4f50-ba83-70589c965a50\" (UID: \"65f8103b-a3c8-4f50-ba83-70589c965a50\") " Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.533981 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65f8103b-a3c8-4f50-ba83-70589c965a50-kube-api-access-bxz9x" (OuterVolumeSpecName: "kube-api-access-bxz9x") pod "65f8103b-a3c8-4f50-ba83-70589c965a50" (UID: "65f8103b-a3c8-4f50-ba83-70589c965a50"). InnerVolumeSpecName "kube-api-access-bxz9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.556277 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65f8103b-a3c8-4f50-ba83-70589c965a50" (UID: "65f8103b-a3c8-4f50-ba83-70589c965a50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.556389 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-config-data" (OuterVolumeSpecName: "config-data") pod "65f8103b-a3c8-4f50-ba83-70589c965a50" (UID: "65f8103b-a3c8-4f50-ba83-70589c965a50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.628594 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.628636 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxz9x\" (UniqueName: \"kubernetes.io/projected/65f8103b-a3c8-4f50-ba83-70589c965a50-kube-api-access-bxz9x\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.628665 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f8103b-a3c8-4f50-ba83-70589c965a50-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.933761 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:23:00 crc kubenswrapper[4944]: I1124 10:23:00.963215 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.043884 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-config-data\") pod \"12e5c569-ab01-4247-8b57-7c3616ac81af\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.044068 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12e5c569-ab01-4247-8b57-7c3616ac81af-logs\") pod \"12e5c569-ab01-4247-8b57-7c3616ac81af\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.044123 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-combined-ca-bundle\") pod \"12e5c569-ab01-4247-8b57-7c3616ac81af\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.044228 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t77hc\" (UniqueName: \"kubernetes.io/projected/12e5c569-ab01-4247-8b57-7c3616ac81af-kube-api-access-t77hc\") pod \"12e5c569-ab01-4247-8b57-7c3616ac81af\" (UID: \"12e5c569-ab01-4247-8b57-7c3616ac81af\") " Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.044757 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12e5c569-ab01-4247-8b57-7c3616ac81af-logs" (OuterVolumeSpecName: "logs") pod "12e5c569-ab01-4247-8b57-7c3616ac81af" (UID: "12e5c569-ab01-4247-8b57-7c3616ac81af"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.045092 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12e5c569-ab01-4247-8b57-7c3616ac81af-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.050515 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12e5c569-ab01-4247-8b57-7c3616ac81af-kube-api-access-t77hc" (OuterVolumeSpecName: "kube-api-access-t77hc") pod "12e5c569-ab01-4247-8b57-7c3616ac81af" (UID: "12e5c569-ab01-4247-8b57-7c3616ac81af"). InnerVolumeSpecName "kube-api-access-t77hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.071095 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-config-data" (OuterVolumeSpecName: "config-data") pod "12e5c569-ab01-4247-8b57-7c3616ac81af" (UID: "12e5c569-ab01-4247-8b57-7c3616ac81af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.074241 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12e5c569-ab01-4247-8b57-7c3616ac81af" (UID: "12e5c569-ab01-4247-8b57-7c3616ac81af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.114684 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"65f8103b-a3c8-4f50-ba83-70589c965a50","Type":"ContainerDied","Data":"cd858913aa5b1c97e35d11b3b37cc89ac948be82ae4a04389302bbb27d2410bf"} Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.114737 4944 scope.go:117] "RemoveContainer" containerID="cf46c94b591c2e8588364e01dc9d6626d8756a4890a25b09c8d19571a5579653" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.114740 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.118275 4944 generic.go:334] "Generic (PLEG): container finished" podID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerID="ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce" exitCode=0 Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.118317 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12e5c569-ab01-4247-8b57-7c3616ac81af","Type":"ContainerDied","Data":"ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce"} Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.118363 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12e5c569-ab01-4247-8b57-7c3616ac81af","Type":"ContainerDied","Data":"e5e5eb89b85f963794ef80a032bb4abde0baf2d0306372a626d4202d5f1b265b"} Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.118938 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.120737 4944 generic.go:334] "Generic (PLEG): container finished" podID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerID="696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd" exitCode=0 Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.120765 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"984463ec-0877-4461-a2a3-7d677c4ea1bb","Type":"ContainerDied","Data":"696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd"} Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.120782 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"984463ec-0877-4461-a2a3-7d677c4ea1bb","Type":"ContainerDied","Data":"919ac8a4773614940a41fc1a857a68114f28082725f0d52627802a19f53ef0d5"} Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.120792 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.146450 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-config-data\") pod \"984463ec-0877-4461-a2a3-7d677c4ea1bb\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.146727 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzpnw\" (UniqueName: \"kubernetes.io/projected/984463ec-0877-4461-a2a3-7d677c4ea1bb-kube-api-access-pzpnw\") pod \"984463ec-0877-4461-a2a3-7d677c4ea1bb\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.146875 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/984463ec-0877-4461-a2a3-7d677c4ea1bb-logs\") pod \"984463ec-0877-4461-a2a3-7d677c4ea1bb\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.147079 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-combined-ca-bundle\") pod \"984463ec-0877-4461-a2a3-7d677c4ea1bb\" (UID: \"984463ec-0877-4461-a2a3-7d677c4ea1bb\") " Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.147371 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/984463ec-0877-4461-a2a3-7d677c4ea1bb-logs" (OuterVolumeSpecName: "logs") pod "984463ec-0877-4461-a2a3-7d677c4ea1bb" (UID: "984463ec-0877-4461-a2a3-7d677c4ea1bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.147841 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.147942 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/984463ec-0877-4461-a2a3-7d677c4ea1bb-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.148017 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e5c569-ab01-4247-8b57-7c3616ac81af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.148149 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t77hc\" (UniqueName: \"kubernetes.io/projected/12e5c569-ab01-4247-8b57-7c3616ac81af-kube-api-access-t77hc\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.152252 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/984463ec-0877-4461-a2a3-7d677c4ea1bb-kube-api-access-pzpnw" (OuterVolumeSpecName: "kube-api-access-pzpnw") pod "984463ec-0877-4461-a2a3-7d677c4ea1bb" (UID: "984463ec-0877-4461-a2a3-7d677c4ea1bb"). InnerVolumeSpecName "kube-api-access-pzpnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.163193 4944 scope.go:117] "RemoveContainer" containerID="ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.175429 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "984463ec-0877-4461-a2a3-7d677c4ea1bb" (UID: "984463ec-0877-4461-a2a3-7d677c4ea1bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.179652 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-config-data" (OuterVolumeSpecName: "config-data") pod "984463ec-0877-4461-a2a3-7d677c4ea1bb" (UID: "984463ec-0877-4461-a2a3-7d677c4ea1bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.250501 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.250544 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzpnw\" (UniqueName: \"kubernetes.io/projected/984463ec-0877-4461-a2a3-7d677c4ea1bb-kube-api-access-pzpnw\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.250555 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984463ec-0877-4461-a2a3-7d677c4ea1bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.252538 4944 scope.go:117] "RemoveContainer" containerID="fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.279621 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.298885 4944 scope.go:117] "RemoveContainer" containerID="ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.305170 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: E1124 10:23:01.306891 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce\": container with ID starting with ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce not found: ID does not exist" containerID="ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.306997 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce"} err="failed to get container status \"ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce\": rpc error: code = NotFound desc = could not find container \"ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce\": container with ID starting with ed8bef7991b99f6aeb00559b9e86353e08c4e68aeaaae451034649616b21d7ce not found: ID does not exist" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.307099 4944 scope.go:117] "RemoveContainer" containerID="fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a" Nov 24 10:23:01 crc kubenswrapper[4944]: E1124 10:23:01.308167 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a\": container with ID starting with fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a not found: ID does not exist" containerID="fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.308214 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a"} err="failed to get container status \"fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a\": rpc error: code = NotFound desc = could not find container \"fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a\": container with ID starting with fcadd0f9b38470ce781a43dfc221ad9d72ccd46614c0ddca4fba7b266eea103a not found: ID does not exist" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.308238 4944 scope.go:117] "RemoveContainer" containerID="696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.333709 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.335707 4944 scope.go:117] "RemoveContainer" containerID="4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.343471 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.351812 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: E1124 10:23:01.352693 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerName="nova-api-api" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.352768 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerName="nova-api-api" Nov 24 10:23:01 crc kubenswrapper[4944]: E1124 10:23:01.352851 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerName="nova-metadata-log" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.352898 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerName="nova-metadata-log" Nov 24 10:23:01 crc kubenswrapper[4944]: E1124 10:23:01.353020 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa3778d-7506-407d-b5d7-b32b310c36d9" containerName="nova-manage" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.353104 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa3778d-7506-407d-b5d7-b32b310c36d9" containerName="nova-manage" Nov 24 10:23:01 crc kubenswrapper[4944]: E1124 10:23:01.353160 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerName="nova-api-log" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.353205 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerName="nova-api-log" Nov 24 10:23:01 crc kubenswrapper[4944]: E1124 10:23:01.353265 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerName="nova-metadata-metadata" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.353314 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerName="nova-metadata-metadata" Nov 24 10:23:01 crc kubenswrapper[4944]: E1124 10:23:01.353371 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f8103b-a3c8-4f50-ba83-70589c965a50" containerName="nova-scheduler-scheduler" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.353421 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f8103b-a3c8-4f50-ba83-70589c965a50" containerName="nova-scheduler-scheduler" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.353666 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa3778d-7506-407d-b5d7-b32b310c36d9" containerName="nova-manage" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.353732 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerName="nova-api-api" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.353783 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerName="nova-metadata-metadata" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.353843 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e5c569-ab01-4247-8b57-7c3616ac81af" containerName="nova-api-log" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.353904 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="984463ec-0877-4461-a2a3-7d677c4ea1bb" containerName="nova-metadata-log" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.353960 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f8103b-a3c8-4f50-ba83-70589c965a50" containerName="nova-scheduler-scheduler" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.355084 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.357468 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.358335 4944 scope.go:117] "RemoveContainer" containerID="696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd" Nov 24 10:23:01 crc kubenswrapper[4944]: E1124 10:23:01.358955 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd\": container with ID starting with 696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd not found: ID does not exist" containerID="696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.358984 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd"} err="failed to get container status \"696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd\": rpc error: code = NotFound desc = could not find container \"696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd\": container with ID starting with 696e682c2d0e66b4ddbe275c652c5ed1eb474045a3d33b187cdc18c25d357ffd not found: ID does not exist" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.359010 4944 scope.go:117] "RemoveContainer" containerID="4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2" Nov 24 10:23:01 crc kubenswrapper[4944]: E1124 10:23:01.359286 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2\": container with ID starting with 4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2 not found: ID does not exist" containerID="4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.359316 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2"} err="failed to get container status \"4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2\": rpc error: code = NotFound desc = could not find container \"4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2\": container with ID starting with 4c02e6eeb99403dfdd2bc3960f5394994c62dc5e61ea2dd681a86874728fd4b2 not found: ID does not exist" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.360697 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.388123 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.391329 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.393945 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.400957 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.454691 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.454899 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.455092 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-config-data\") pod \"nova-scheduler-0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.455155 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4dv9\" (UniqueName: \"kubernetes.io/projected/497cebe3-1e6d-4294-bef1-45453ace62f0-kube-api-access-t4dv9\") pod \"nova-scheduler-0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.472343 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.485994 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.487641 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.492483 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.497223 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.557170 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/401b7639-8818-40f5-9c89-cbeacfcf465a-logs\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.557236 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.557272 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.557324 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-config-data\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.557357 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-config-data\") pod \"nova-scheduler-0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.557396 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf6vs\" (UniqueName: \"kubernetes.io/projected/401b7639-8818-40f5-9c89-cbeacfcf465a-kube-api-access-hf6vs\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.557422 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4dv9\" (UniqueName: \"kubernetes.io/projected/497cebe3-1e6d-4294-bef1-45453ace62f0-kube-api-access-t4dv9\") pod \"nova-scheduler-0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.562444 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.564143 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-config-data\") pod \"nova-scheduler-0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.574678 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4dv9\" (UniqueName: \"kubernetes.io/projected/497cebe3-1e6d-4294-bef1-45453ace62f0-kube-api-access-t4dv9\") pod \"nova-scheduler-0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.659259 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-config-data\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.659339 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8dz9\" (UniqueName: \"kubernetes.io/projected/3b8d3337-080c-4f51-be4d-9a69f6a6e074-kube-api-access-p8dz9\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.659372 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf6vs\" (UniqueName: \"kubernetes.io/projected/401b7639-8818-40f5-9c89-cbeacfcf465a-kube-api-access-hf6vs\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.659913 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/401b7639-8818-40f5-9c89-cbeacfcf465a-logs\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.660473 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.660507 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.660557 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-config-data\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.660597 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b8d3337-080c-4f51-be4d-9a69f6a6e074-logs\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.660434 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/401b7639-8818-40f5-9c89-cbeacfcf465a-logs\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.664857 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.666166 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-config-data\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.674969 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf6vs\" (UniqueName: \"kubernetes.io/projected/401b7639-8818-40f5-9c89-cbeacfcf465a-kube-api-access-hf6vs\") pod \"nova-api-0\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.680974 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.712962 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.761809 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.761913 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b8d3337-080c-4f51-be4d-9a69f6a6e074-logs\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.761941 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-config-data\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.761965 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8dz9\" (UniqueName: \"kubernetes.io/projected/3b8d3337-080c-4f51-be4d-9a69f6a6e074-kube-api-access-p8dz9\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.762685 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b8d3337-080c-4f51-be4d-9a69f6a6e074-logs\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.769813 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.772383 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-config-data\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.785930 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8dz9\" (UniqueName: \"kubernetes.io/projected/3b8d3337-080c-4f51-be4d-9a69f6a6e074-kube-api-access-p8dz9\") pod \"nova-metadata-0\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " pod="openstack/nova-metadata-0" Nov 24 10:23:01 crc kubenswrapper[4944]: I1124 10:23:01.811396 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:23:02 crc kubenswrapper[4944]: I1124 10:23:02.108742 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:23:02 crc kubenswrapper[4944]: W1124 10:23:02.114635 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod497cebe3_1e6d_4294_bef1_45453ace62f0.slice/crio-d58d58a7748545c88ed14f5e7f8d5204d62af14bc6c13975e1407a289f5bdff0 WatchSource:0}: Error finding container d58d58a7748545c88ed14f5e7f8d5204d62af14bc6c13975e1407a289f5bdff0: Status 404 returned error can't find the container with id d58d58a7748545c88ed14f5e7f8d5204d62af14bc6c13975e1407a289f5bdff0 Nov 24 10:23:02 crc kubenswrapper[4944]: I1124 10:23:02.130235 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"497cebe3-1e6d-4294-bef1-45453ace62f0","Type":"ContainerStarted","Data":"d58d58a7748545c88ed14f5e7f8d5204d62af14bc6c13975e1407a289f5bdff0"} Nov 24 10:23:02 crc kubenswrapper[4944]: I1124 10:23:02.198734 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:23:02 crc kubenswrapper[4944]: W1124 10:23:02.284960 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b8d3337_080c_4f51_be4d_9a69f6a6e074.slice/crio-a526a1fdf2b77de9aa2e9601dead235ea4aecb9d51497a4ab42ef52dec1453fe WatchSource:0}: Error finding container a526a1fdf2b77de9aa2e9601dead235ea4aecb9d51497a4ab42ef52dec1453fe: Status 404 returned error can't find the container with id a526a1fdf2b77de9aa2e9601dead235ea4aecb9d51497a4ab42ef52dec1453fe Nov 24 10:23:02 crc kubenswrapper[4944]: I1124 10:23:02.290814 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12e5c569-ab01-4247-8b57-7c3616ac81af" path="/var/lib/kubelet/pods/12e5c569-ab01-4247-8b57-7c3616ac81af/volumes" Nov 24 10:23:02 crc kubenswrapper[4944]: I1124 10:23:02.291756 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65f8103b-a3c8-4f50-ba83-70589c965a50" path="/var/lib/kubelet/pods/65f8103b-a3c8-4f50-ba83-70589c965a50/volumes" Nov 24 10:23:02 crc kubenswrapper[4944]: I1124 10:23:02.292431 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="984463ec-0877-4461-a2a3-7d677c4ea1bb" path="/var/lib/kubelet/pods/984463ec-0877-4461-a2a3-7d677c4ea1bb/volumes" Nov 24 10:23:02 crc kubenswrapper[4944]: I1124 10:23:02.293493 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:23:03 crc kubenswrapper[4944]: I1124 10:23:03.142539 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"497cebe3-1e6d-4294-bef1-45453ace62f0","Type":"ContainerStarted","Data":"3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f"} Nov 24 10:23:03 crc kubenswrapper[4944]: I1124 10:23:03.145235 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"401b7639-8818-40f5-9c89-cbeacfcf465a","Type":"ContainerStarted","Data":"58bc6140c220ada0743bf577e4ee73036b60a7bf502be440a2e8a7539fa4abf5"} Nov 24 10:23:03 crc kubenswrapper[4944]: I1124 10:23:03.145368 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"401b7639-8818-40f5-9c89-cbeacfcf465a","Type":"ContainerStarted","Data":"0d5e1d6812a5dcca4341f23087431fa85f7b0889b653a4012082927656c833b6"} Nov 24 10:23:03 crc kubenswrapper[4944]: I1124 10:23:03.145383 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"401b7639-8818-40f5-9c89-cbeacfcf465a","Type":"ContainerStarted","Data":"684b55fc4bfaf8641f206857eb3d5754a746e35b95b5ca3a2a8db846aec74a0f"} Nov 24 10:23:03 crc kubenswrapper[4944]: I1124 10:23:03.147189 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b8d3337-080c-4f51-be4d-9a69f6a6e074","Type":"ContainerStarted","Data":"fbb488a6adfa0c34cce9196736c83a62353ec6209398f2fea648fb462c89a43b"} Nov 24 10:23:03 crc kubenswrapper[4944]: I1124 10:23:03.147244 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b8d3337-080c-4f51-be4d-9a69f6a6e074","Type":"ContainerStarted","Data":"a79308575fa58b1f06e8a3c2da13f2e44ce9fa9b5031db96d89f6ca9517c741f"} Nov 24 10:23:03 crc kubenswrapper[4944]: I1124 10:23:03.147257 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b8d3337-080c-4f51-be4d-9a69f6a6e074","Type":"ContainerStarted","Data":"a526a1fdf2b77de9aa2e9601dead235ea4aecb9d51497a4ab42ef52dec1453fe"} Nov 24 10:23:03 crc kubenswrapper[4944]: I1124 10:23:03.165993 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.165968647 podStartE2EDuration="2.165968647s" podCreationTimestamp="2025-11-24 10:23:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:23:03.161091232 +0000 UTC m=+5443.695531714" watchObservedRunningTime="2025-11-24 10:23:03.165968647 +0000 UTC m=+5443.700409109" Nov 24 10:23:03 crc kubenswrapper[4944]: I1124 10:23:03.192710 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.192686917 podStartE2EDuration="2.192686917s" podCreationTimestamp="2025-11-24 10:23:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:23:03.187216203 +0000 UTC m=+5443.721656665" watchObservedRunningTime="2025-11-24 10:23:03.192686917 +0000 UTC m=+5443.727127379" Nov 24 10:23:06 crc kubenswrapper[4944]: I1124 10:23:06.681990 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 10:23:06 crc kubenswrapper[4944]: I1124 10:23:06.812120 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 10:23:06 crc kubenswrapper[4944]: I1124 10:23:06.812203 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 10:23:11 crc kubenswrapper[4944]: I1124 10:23:11.681665 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 10:23:11 crc kubenswrapper[4944]: I1124 10:23:11.709624 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 10:23:11 crc kubenswrapper[4944]: I1124 10:23:11.713610 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 10:23:11 crc kubenswrapper[4944]: I1124 10:23:11.713655 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 10:23:11 crc kubenswrapper[4944]: I1124 10:23:11.744001 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=10.743986276 podStartE2EDuration="10.743986276s" podCreationTimestamp="2025-11-24 10:23:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:23:03.214014877 +0000 UTC m=+5443.748455339" watchObservedRunningTime="2025-11-24 10:23:11.743986276 +0000 UTC m=+5452.278426738" Nov 24 10:23:11 crc kubenswrapper[4944]: I1124 10:23:11.812982 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 10:23:11 crc kubenswrapper[4944]: I1124 10:23:11.813054 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 10:23:12 crc kubenswrapper[4944]: I1124 10:23:12.246913 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 10:23:12 crc kubenswrapper[4944]: I1124 10:23:12.796238 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.72:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:23:12 crc kubenswrapper[4944]: I1124 10:23:12.796287 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.72:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:23:12 crc kubenswrapper[4944]: I1124 10:23:12.898300 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.73:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:23:12 crc kubenswrapper[4944]: I1124 10:23:12.898300 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.73:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.716861 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.717495 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.717879 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.717899 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.720835 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.722338 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.822448 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.824552 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.826388 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.901133 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-574d65f5bf-8sw2n"] Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.902640 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:21 crc kubenswrapper[4944]: I1124 10:23:21.915446 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-574d65f5bf-8sw2n"] Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.018241 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-dns-svc\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.018283 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckslw\" (UniqueName: \"kubernetes.io/projected/a797b85c-0a33-4a54-a3cd-245b74bacba4-kube-api-access-ckslw\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.018332 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-nb\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.018584 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-sb\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.018741 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-config\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.121039 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-sb\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.121147 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-config\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.121190 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-dns-svc\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.121213 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckslw\" (UniqueName: \"kubernetes.io/projected/a797b85c-0a33-4a54-a3cd-245b74bacba4-kube-api-access-ckslw\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.121268 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-nb\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.122146 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-sb\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.122262 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-config\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.122278 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-nb\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.122387 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-dns-svc\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.142173 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckslw\" (UniqueName: \"kubernetes.io/projected/a797b85c-0a33-4a54-a3cd-245b74bacba4-kube-api-access-ckslw\") pod \"dnsmasq-dns-574d65f5bf-8sw2n\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.233843 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.309739 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 10:23:22 crc kubenswrapper[4944]: I1124 10:23:22.747029 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-574d65f5bf-8sw2n"] Nov 24 10:23:23 crc kubenswrapper[4944]: I1124 10:23:23.311543 4944 generic.go:334] "Generic (PLEG): container finished" podID="a797b85c-0a33-4a54-a3cd-245b74bacba4" containerID="af3f54a378fe3878876ef46b3ad3b36dc93cd8e9db0901ca8eca74b0f84e213d" exitCode=0 Nov 24 10:23:23 crc kubenswrapper[4944]: I1124 10:23:23.311655 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" event={"ID":"a797b85c-0a33-4a54-a3cd-245b74bacba4","Type":"ContainerDied","Data":"af3f54a378fe3878876ef46b3ad3b36dc93cd8e9db0901ca8eca74b0f84e213d"} Nov 24 10:23:23 crc kubenswrapper[4944]: I1124 10:23:23.311908 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" event={"ID":"a797b85c-0a33-4a54-a3cd-245b74bacba4","Type":"ContainerStarted","Data":"e146f726d62b3c8ec6ae2b0f2ba798b8326b5eaa927dbd6e1df07f3fb8f096aa"} Nov 24 10:23:23 crc kubenswrapper[4944]: I1124 10:23:23.548463 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:23:23 crc kubenswrapper[4944]: I1124 10:23:23.548525 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:23:24 crc kubenswrapper[4944]: I1124 10:23:24.324705 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" event={"ID":"a797b85c-0a33-4a54-a3cd-245b74bacba4","Type":"ContainerStarted","Data":"68556ce24ecb94075a6493bb376a6960468563d64756480c2369c11ba32382b9"} Nov 24 10:23:24 crc kubenswrapper[4944]: I1124 10:23:24.325117 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:24 crc kubenswrapper[4944]: I1124 10:23:24.354921 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" podStartSLOduration=3.354896879 podStartE2EDuration="3.354896879s" podCreationTimestamp="2025-11-24 10:23:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:23:24.351111479 +0000 UTC m=+5464.885551981" watchObservedRunningTime="2025-11-24 10:23:24.354896879 +0000 UTC m=+5464.889337341" Nov 24 10:23:30 crc kubenswrapper[4944]: I1124 10:23:30.965518 4944 scope.go:117] "RemoveContainer" containerID="cd8c44c5fe857a74c32862982e3de75c9f01328b42d198a0fc18363d5ac5b247" Nov 24 10:23:32 crc kubenswrapper[4944]: I1124 10:23:32.235295 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:32 crc kubenswrapper[4944]: I1124 10:23:32.604107 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b58cb654c-gwclv"] Nov 24 10:23:32 crc kubenswrapper[4944]: I1124 10:23:32.604677 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" podUID="ea6d6441-ea1d-43df-b4dc-6b838f736b05" containerName="dnsmasq-dns" containerID="cri-o://70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff" gracePeriod=10 Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.151512 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.252505 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mlpf\" (UniqueName: \"kubernetes.io/projected/ea6d6441-ea1d-43df-b4dc-6b838f736b05-kube-api-access-6mlpf\") pod \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.253316 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-sb\") pod \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.253383 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-config\") pod \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.253535 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-nb\") pod \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.253691 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-dns-svc\") pod \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\" (UID: \"ea6d6441-ea1d-43df-b4dc-6b838f736b05\") " Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.262118 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea6d6441-ea1d-43df-b4dc-6b838f736b05-kube-api-access-6mlpf" (OuterVolumeSpecName: "kube-api-access-6mlpf") pod "ea6d6441-ea1d-43df-b4dc-6b838f736b05" (UID: "ea6d6441-ea1d-43df-b4dc-6b838f736b05"). InnerVolumeSpecName "kube-api-access-6mlpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.304291 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-config" (OuterVolumeSpecName: "config") pod "ea6d6441-ea1d-43df-b4dc-6b838f736b05" (UID: "ea6d6441-ea1d-43df-b4dc-6b838f736b05"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.320125 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea6d6441-ea1d-43df-b4dc-6b838f736b05" (UID: "ea6d6441-ea1d-43df-b4dc-6b838f736b05"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.320622 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea6d6441-ea1d-43df-b4dc-6b838f736b05" (UID: "ea6d6441-ea1d-43df-b4dc-6b838f736b05"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.332187 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea6d6441-ea1d-43df-b4dc-6b838f736b05" (UID: "ea6d6441-ea1d-43df-b4dc-6b838f736b05"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.356891 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.356927 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mlpf\" (UniqueName: \"kubernetes.io/projected/ea6d6441-ea1d-43df-b4dc-6b838f736b05-kube-api-access-6mlpf\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.356945 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.356959 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.356971 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea6d6441-ea1d-43df-b4dc-6b838f736b05-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.569397 4944 generic.go:334] "Generic (PLEG): container finished" podID="ea6d6441-ea1d-43df-b4dc-6b838f736b05" containerID="70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff" exitCode=0 Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.569449 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" event={"ID":"ea6d6441-ea1d-43df-b4dc-6b838f736b05","Type":"ContainerDied","Data":"70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff"} Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.569478 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" event={"ID":"ea6d6441-ea1d-43df-b4dc-6b838f736b05","Type":"ContainerDied","Data":"53b5de04b868cf32b9c9f697180533ee5d36e8e518c0c281da23fb092fc5c2ba"} Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.569499 4944 scope.go:117] "RemoveContainer" containerID="70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.569816 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b58cb654c-gwclv" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.596144 4944 scope.go:117] "RemoveContainer" containerID="5eff8c1c55bd1a723ba72052f7514edbe5625d7d219e61e1046cbc0f56bfc299" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.607169 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b58cb654c-gwclv"] Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.617278 4944 scope.go:117] "RemoveContainer" containerID="70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.617459 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b58cb654c-gwclv"] Nov 24 10:23:33 crc kubenswrapper[4944]: E1124 10:23:33.617708 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff\": container with ID starting with 70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff not found: ID does not exist" containerID="70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.617752 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff"} err="failed to get container status \"70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff\": rpc error: code = NotFound desc = could not find container \"70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff\": container with ID starting with 70c2c85dc9443b95b3eb160afa59a9272c465a2c00484eb41bb8d156581536ff not found: ID does not exist" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.617787 4944 scope.go:117] "RemoveContainer" containerID="5eff8c1c55bd1a723ba72052f7514edbe5625d7d219e61e1046cbc0f56bfc299" Nov 24 10:23:33 crc kubenswrapper[4944]: E1124 10:23:33.618205 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eff8c1c55bd1a723ba72052f7514edbe5625d7d219e61e1046cbc0f56bfc299\": container with ID starting with 5eff8c1c55bd1a723ba72052f7514edbe5625d7d219e61e1046cbc0f56bfc299 not found: ID does not exist" containerID="5eff8c1c55bd1a723ba72052f7514edbe5625d7d219e61e1046cbc0f56bfc299" Nov 24 10:23:33 crc kubenswrapper[4944]: I1124 10:23:33.618527 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eff8c1c55bd1a723ba72052f7514edbe5625d7d219e61e1046cbc0f56bfc299"} err="failed to get container status \"5eff8c1c55bd1a723ba72052f7514edbe5625d7d219e61e1046cbc0f56bfc299\": rpc error: code = NotFound desc = could not find container \"5eff8c1c55bd1a723ba72052f7514edbe5625d7d219e61e1046cbc0f56bfc299\": container with ID starting with 5eff8c1c55bd1a723ba72052f7514edbe5625d7d219e61e1046cbc0f56bfc299 not found: ID does not exist" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.050278 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-m29ll"] Nov 24 10:23:34 crc kubenswrapper[4944]: E1124 10:23:34.050749 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea6d6441-ea1d-43df-b4dc-6b838f736b05" containerName="dnsmasq-dns" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.050773 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea6d6441-ea1d-43df-b4dc-6b838f736b05" containerName="dnsmasq-dns" Nov 24 10:23:34 crc kubenswrapper[4944]: E1124 10:23:34.050815 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea6d6441-ea1d-43df-b4dc-6b838f736b05" containerName="init" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.050824 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea6d6441-ea1d-43df-b4dc-6b838f736b05" containerName="init" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.051033 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea6d6441-ea1d-43df-b4dc-6b838f736b05" containerName="dnsmasq-dns" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.051982 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m29ll" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.060497 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-9dae-account-create-xbfr7"] Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.061784 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9dae-account-create-xbfr7" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.072641 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.075475 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9dae-account-create-xbfr7"] Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.084724 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-m29ll"] Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.175186 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d028bc5d-4e09-41eb-bf7b-78de5e065799-operator-scripts\") pod \"cinder-9dae-account-create-xbfr7\" (UID: \"d028bc5d-4e09-41eb-bf7b-78de5e065799\") " pod="openstack/cinder-9dae-account-create-xbfr7" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.175394 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4j44\" (UniqueName: \"kubernetes.io/projected/d028bc5d-4e09-41eb-bf7b-78de5e065799-kube-api-access-n4j44\") pod \"cinder-9dae-account-create-xbfr7\" (UID: \"d028bc5d-4e09-41eb-bf7b-78de5e065799\") " pod="openstack/cinder-9dae-account-create-xbfr7" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.175564 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e93583a-c60e-4665-8b4a-aad77baf1ac6-operator-scripts\") pod \"cinder-db-create-m29ll\" (UID: \"5e93583a-c60e-4665-8b4a-aad77baf1ac6\") " pod="openstack/cinder-db-create-m29ll" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.175704 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcktd\" (UniqueName: \"kubernetes.io/projected/5e93583a-c60e-4665-8b4a-aad77baf1ac6-kube-api-access-rcktd\") pod \"cinder-db-create-m29ll\" (UID: \"5e93583a-c60e-4665-8b4a-aad77baf1ac6\") " pod="openstack/cinder-db-create-m29ll" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.277345 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcktd\" (UniqueName: \"kubernetes.io/projected/5e93583a-c60e-4665-8b4a-aad77baf1ac6-kube-api-access-rcktd\") pod \"cinder-db-create-m29ll\" (UID: \"5e93583a-c60e-4665-8b4a-aad77baf1ac6\") " pod="openstack/cinder-db-create-m29ll" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.277903 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d028bc5d-4e09-41eb-bf7b-78de5e065799-operator-scripts\") pod \"cinder-9dae-account-create-xbfr7\" (UID: \"d028bc5d-4e09-41eb-bf7b-78de5e065799\") " pod="openstack/cinder-9dae-account-create-xbfr7" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.278035 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4j44\" (UniqueName: \"kubernetes.io/projected/d028bc5d-4e09-41eb-bf7b-78de5e065799-kube-api-access-n4j44\") pod \"cinder-9dae-account-create-xbfr7\" (UID: \"d028bc5d-4e09-41eb-bf7b-78de5e065799\") " pod="openstack/cinder-9dae-account-create-xbfr7" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.278246 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e93583a-c60e-4665-8b4a-aad77baf1ac6-operator-scripts\") pod \"cinder-db-create-m29ll\" (UID: \"5e93583a-c60e-4665-8b4a-aad77baf1ac6\") " pod="openstack/cinder-db-create-m29ll" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.279080 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d028bc5d-4e09-41eb-bf7b-78de5e065799-operator-scripts\") pod \"cinder-9dae-account-create-xbfr7\" (UID: \"d028bc5d-4e09-41eb-bf7b-78de5e065799\") " pod="openstack/cinder-9dae-account-create-xbfr7" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.279095 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e93583a-c60e-4665-8b4a-aad77baf1ac6-operator-scripts\") pod \"cinder-db-create-m29ll\" (UID: \"5e93583a-c60e-4665-8b4a-aad77baf1ac6\") " pod="openstack/cinder-db-create-m29ll" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.288128 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea6d6441-ea1d-43df-b4dc-6b838f736b05" path="/var/lib/kubelet/pods/ea6d6441-ea1d-43df-b4dc-6b838f736b05/volumes" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.295533 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcktd\" (UniqueName: \"kubernetes.io/projected/5e93583a-c60e-4665-8b4a-aad77baf1ac6-kube-api-access-rcktd\") pod \"cinder-db-create-m29ll\" (UID: \"5e93583a-c60e-4665-8b4a-aad77baf1ac6\") " pod="openstack/cinder-db-create-m29ll" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.297365 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4j44\" (UniqueName: \"kubernetes.io/projected/d028bc5d-4e09-41eb-bf7b-78de5e065799-kube-api-access-n4j44\") pod \"cinder-9dae-account-create-xbfr7\" (UID: \"d028bc5d-4e09-41eb-bf7b-78de5e065799\") " pod="openstack/cinder-9dae-account-create-xbfr7" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.374374 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m29ll" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.385036 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9dae-account-create-xbfr7" Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.849305 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-m29ll"] Nov 24 10:23:34 crc kubenswrapper[4944]: W1124 10:23:34.851416 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e93583a_c60e_4665_8b4a_aad77baf1ac6.slice/crio-5b829858efeabb12468eacfb15a2db35ba1c6f2fd0e4d6068a20624de3a0e19b WatchSource:0}: Error finding container 5b829858efeabb12468eacfb15a2db35ba1c6f2fd0e4d6068a20624de3a0e19b: Status 404 returned error can't find the container with id 5b829858efeabb12468eacfb15a2db35ba1c6f2fd0e4d6068a20624de3a0e19b Nov 24 10:23:34 crc kubenswrapper[4944]: I1124 10:23:34.940496 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9dae-account-create-xbfr7"] Nov 24 10:23:34 crc kubenswrapper[4944]: W1124 10:23:34.955916 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd028bc5d_4e09_41eb_bf7b_78de5e065799.slice/crio-4e0f1cf1be31e71c14e83050ea7d0a29fb89a87b3631a3cae6d6ff4b7811feb3 WatchSource:0}: Error finding container 4e0f1cf1be31e71c14e83050ea7d0a29fb89a87b3631a3cae6d6ff4b7811feb3: Status 404 returned error can't find the container with id 4e0f1cf1be31e71c14e83050ea7d0a29fb89a87b3631a3cae6d6ff4b7811feb3 Nov 24 10:23:35 crc kubenswrapper[4944]: I1124 10:23:35.592132 4944 generic.go:334] "Generic (PLEG): container finished" podID="5e93583a-c60e-4665-8b4a-aad77baf1ac6" containerID="b6f5c3afeedac08662dbac9c37ec2608bf5c9543bc0d6d9cabaa50a78e72513a" exitCode=0 Nov 24 10:23:35 crc kubenswrapper[4944]: I1124 10:23:35.592199 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-m29ll" event={"ID":"5e93583a-c60e-4665-8b4a-aad77baf1ac6","Type":"ContainerDied","Data":"b6f5c3afeedac08662dbac9c37ec2608bf5c9543bc0d6d9cabaa50a78e72513a"} Nov 24 10:23:35 crc kubenswrapper[4944]: I1124 10:23:35.592710 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-m29ll" event={"ID":"5e93583a-c60e-4665-8b4a-aad77baf1ac6","Type":"ContainerStarted","Data":"5b829858efeabb12468eacfb15a2db35ba1c6f2fd0e4d6068a20624de3a0e19b"} Nov 24 10:23:35 crc kubenswrapper[4944]: I1124 10:23:35.595498 4944 generic.go:334] "Generic (PLEG): container finished" podID="d028bc5d-4e09-41eb-bf7b-78de5e065799" containerID="cd8c79a09e89d838bd4075b59169a50473fdf8071cd1897ceeaf005e16b8f143" exitCode=0 Nov 24 10:23:35 crc kubenswrapper[4944]: I1124 10:23:35.595546 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9dae-account-create-xbfr7" event={"ID":"d028bc5d-4e09-41eb-bf7b-78de5e065799","Type":"ContainerDied","Data":"cd8c79a09e89d838bd4075b59169a50473fdf8071cd1897ceeaf005e16b8f143"} Nov 24 10:23:35 crc kubenswrapper[4944]: I1124 10:23:35.595614 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9dae-account-create-xbfr7" event={"ID":"d028bc5d-4e09-41eb-bf7b-78de5e065799","Type":"ContainerStarted","Data":"4e0f1cf1be31e71c14e83050ea7d0a29fb89a87b3631a3cae6d6ff4b7811feb3"} Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:36.999890 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9dae-account-create-xbfr7" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.006184 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m29ll" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.134248 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4j44\" (UniqueName: \"kubernetes.io/projected/d028bc5d-4e09-41eb-bf7b-78de5e065799-kube-api-access-n4j44\") pod \"d028bc5d-4e09-41eb-bf7b-78de5e065799\" (UID: \"d028bc5d-4e09-41eb-bf7b-78de5e065799\") " Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.134355 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e93583a-c60e-4665-8b4a-aad77baf1ac6-operator-scripts\") pod \"5e93583a-c60e-4665-8b4a-aad77baf1ac6\" (UID: \"5e93583a-c60e-4665-8b4a-aad77baf1ac6\") " Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.134455 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcktd\" (UniqueName: \"kubernetes.io/projected/5e93583a-c60e-4665-8b4a-aad77baf1ac6-kube-api-access-rcktd\") pod \"5e93583a-c60e-4665-8b4a-aad77baf1ac6\" (UID: \"5e93583a-c60e-4665-8b4a-aad77baf1ac6\") " Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.134545 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d028bc5d-4e09-41eb-bf7b-78de5e065799-operator-scripts\") pod \"d028bc5d-4e09-41eb-bf7b-78de5e065799\" (UID: \"d028bc5d-4e09-41eb-bf7b-78de5e065799\") " Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.134971 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e93583a-c60e-4665-8b4a-aad77baf1ac6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5e93583a-c60e-4665-8b4a-aad77baf1ac6" (UID: "5e93583a-c60e-4665-8b4a-aad77baf1ac6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.135170 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d028bc5d-4e09-41eb-bf7b-78de5e065799-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d028bc5d-4e09-41eb-bf7b-78de5e065799" (UID: "d028bc5d-4e09-41eb-bf7b-78de5e065799"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.140069 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e93583a-c60e-4665-8b4a-aad77baf1ac6-kube-api-access-rcktd" (OuterVolumeSpecName: "kube-api-access-rcktd") pod "5e93583a-c60e-4665-8b4a-aad77baf1ac6" (UID: "5e93583a-c60e-4665-8b4a-aad77baf1ac6"). InnerVolumeSpecName "kube-api-access-rcktd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.140132 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d028bc5d-4e09-41eb-bf7b-78de5e065799-kube-api-access-n4j44" (OuterVolumeSpecName: "kube-api-access-n4j44") pod "d028bc5d-4e09-41eb-bf7b-78de5e065799" (UID: "d028bc5d-4e09-41eb-bf7b-78de5e065799"). InnerVolumeSpecName "kube-api-access-n4j44". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.236275 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e93583a-c60e-4665-8b4a-aad77baf1ac6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.236316 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcktd\" (UniqueName: \"kubernetes.io/projected/5e93583a-c60e-4665-8b4a-aad77baf1ac6-kube-api-access-rcktd\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.236330 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d028bc5d-4e09-41eb-bf7b-78de5e065799-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.236343 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4j44\" (UniqueName: \"kubernetes.io/projected/d028bc5d-4e09-41eb-bf7b-78de5e065799-kube-api-access-n4j44\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.614672 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9dae-account-create-xbfr7" event={"ID":"d028bc5d-4e09-41eb-bf7b-78de5e065799","Type":"ContainerDied","Data":"4e0f1cf1be31e71c14e83050ea7d0a29fb89a87b3631a3cae6d6ff4b7811feb3"} Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.614718 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e0f1cf1be31e71c14e83050ea7d0a29fb89a87b3631a3cae6d6ff4b7811feb3" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.614695 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9dae-account-create-xbfr7" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.616225 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-m29ll" event={"ID":"5e93583a-c60e-4665-8b4a-aad77baf1ac6","Type":"ContainerDied","Data":"5b829858efeabb12468eacfb15a2db35ba1c6f2fd0e4d6068a20624de3a0e19b"} Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.616244 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b829858efeabb12468eacfb15a2db35ba1c6f2fd0e4d6068a20624de3a0e19b" Nov 24 10:23:37 crc kubenswrapper[4944]: I1124 10:23:37.616289 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-m29ll" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.409737 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-kczbq"] Nov 24 10:23:39 crc kubenswrapper[4944]: E1124 10:23:39.410478 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d028bc5d-4e09-41eb-bf7b-78de5e065799" containerName="mariadb-account-create" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.410491 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d028bc5d-4e09-41eb-bf7b-78de5e065799" containerName="mariadb-account-create" Nov 24 10:23:39 crc kubenswrapper[4944]: E1124 10:23:39.410508 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e93583a-c60e-4665-8b4a-aad77baf1ac6" containerName="mariadb-database-create" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.410515 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e93583a-c60e-4665-8b4a-aad77baf1ac6" containerName="mariadb-database-create" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.410718 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e93583a-c60e-4665-8b4a-aad77baf1ac6" containerName="mariadb-database-create" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.410733 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d028bc5d-4e09-41eb-bf7b-78de5e065799" containerName="mariadb-account-create" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.411391 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.414060 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.414170 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.414246 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-gx6fl" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.419140 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-kczbq"] Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.578425 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-db-sync-config-data\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.578745 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdp5p\" (UniqueName: \"kubernetes.io/projected/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-kube-api-access-sdp5p\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.579165 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-combined-ca-bundle\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.579329 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-scripts\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.579367 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-etc-machine-id\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.579468 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-config-data\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.680971 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-config-data\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.681087 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-db-sync-config-data\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.681141 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdp5p\" (UniqueName: \"kubernetes.io/projected/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-kube-api-access-sdp5p\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.681183 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-combined-ca-bundle\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.681212 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-scripts\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.681247 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-etc-machine-id\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.681358 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-etc-machine-id\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.689087 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-combined-ca-bundle\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.691341 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-config-data\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.692507 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-db-sync-config-data\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.696542 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-scripts\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.715779 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdp5p\" (UniqueName: \"kubernetes.io/projected/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-kube-api-access-sdp5p\") pod \"cinder-db-sync-kczbq\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:39 crc kubenswrapper[4944]: I1124 10:23:39.733211 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:40 crc kubenswrapper[4944]: W1124 10:23:40.296577 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ed874ef_a00f_47b9_a9d8_18e2a8836a8b.slice/crio-a694dc752669b71048975b1fd85f404a518203f9753ce5f3be36995be57e4a7e WatchSource:0}: Error finding container a694dc752669b71048975b1fd85f404a518203f9753ce5f3be36995be57e4a7e: Status 404 returned error can't find the container with id a694dc752669b71048975b1fd85f404a518203f9753ce5f3be36995be57e4a7e Nov 24 10:23:40 crc kubenswrapper[4944]: I1124 10:23:40.299690 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-kczbq"] Nov 24 10:23:40 crc kubenswrapper[4944]: I1124 10:23:40.654637 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kczbq" event={"ID":"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b","Type":"ContainerStarted","Data":"a694dc752669b71048975b1fd85f404a518203f9753ce5f3be36995be57e4a7e"} Nov 24 10:23:41 crc kubenswrapper[4944]: I1124 10:23:41.691726 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kczbq" event={"ID":"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b","Type":"ContainerStarted","Data":"c22915dd38e7aa912865e171d32307a63f691938a23c2eb7f333d8f873ff2379"} Nov 24 10:23:41 crc kubenswrapper[4944]: I1124 10:23:41.712725 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-kczbq" podStartSLOduration=2.712705873 podStartE2EDuration="2.712705873s" podCreationTimestamp="2025-11-24 10:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:23:41.709978606 +0000 UTC m=+5482.244419088" watchObservedRunningTime="2025-11-24 10:23:41.712705873 +0000 UTC m=+5482.247146335" Nov 24 10:23:43 crc kubenswrapper[4944]: I1124 10:23:43.710122 4944 generic.go:334] "Generic (PLEG): container finished" podID="0ed874ef-a00f-47b9-a9d8-18e2a8836a8b" containerID="c22915dd38e7aa912865e171d32307a63f691938a23c2eb7f333d8f873ff2379" exitCode=0 Nov 24 10:23:43 crc kubenswrapper[4944]: I1124 10:23:43.710177 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kczbq" event={"ID":"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b","Type":"ContainerDied","Data":"c22915dd38e7aa912865e171d32307a63f691938a23c2eb7f333d8f873ff2379"} Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.032017 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.186099 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-config-data\") pod \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.186184 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-combined-ca-bundle\") pod \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.186232 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdp5p\" (UniqueName: \"kubernetes.io/projected/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-kube-api-access-sdp5p\") pod \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.186317 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-scripts\") pod \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.186335 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-etc-machine-id\") pod \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.186373 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-db-sync-config-data\") pod \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\" (UID: \"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b\") " Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.186597 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b" (UID: "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.186766 4944 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.192856 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-scripts" (OuterVolumeSpecName: "scripts") pod "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b" (UID: "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.192888 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b" (UID: "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.193102 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-kube-api-access-sdp5p" (OuterVolumeSpecName: "kube-api-access-sdp5p") pod "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b" (UID: "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b"). InnerVolumeSpecName "kube-api-access-sdp5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.215036 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b" (UID: "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.244525 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-config-data" (OuterVolumeSpecName: "config-data") pod "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b" (UID: "0ed874ef-a00f-47b9-a9d8-18e2a8836a8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.288426 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.288453 4944 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.288463 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.288472 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.288479 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdp5p\" (UniqueName: \"kubernetes.io/projected/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b-kube-api-access-sdp5p\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.729469 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kczbq" event={"ID":"0ed874ef-a00f-47b9-a9d8-18e2a8836a8b","Type":"ContainerDied","Data":"a694dc752669b71048975b1fd85f404a518203f9753ce5f3be36995be57e4a7e"} Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.729800 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a694dc752669b71048975b1fd85f404a518203f9753ce5f3be36995be57e4a7e" Nov 24 10:23:45 crc kubenswrapper[4944]: I1124 10:23:45.729713 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kczbq" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.048623 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b4c94945c-2ck29"] Nov 24 10:23:46 crc kubenswrapper[4944]: E1124 10:23:46.050005 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ed874ef-a00f-47b9-a9d8-18e2a8836a8b" containerName="cinder-db-sync" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.050088 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ed874ef-a00f-47b9-a9d8-18e2a8836a8b" containerName="cinder-db-sync" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.050351 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ed874ef-a00f-47b9-a9d8-18e2a8836a8b" containerName="cinder-db-sync" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.051363 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.075181 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b4c94945c-2ck29"] Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.185400 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.188967 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.193125 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.193778 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.194117 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.194321 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-gx6fl" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.211123 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-dns-svc\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.211282 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjs5d\" (UniqueName: \"kubernetes.io/projected/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-kube-api-access-gjs5d\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.211320 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-nb\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.211357 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-sb\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.211422 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-config\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.220016 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.313013 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjs5d\" (UniqueName: \"kubernetes.io/projected/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-kube-api-access-gjs5d\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.313074 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-nb\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.313096 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-sb\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.313117 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qthvq\" (UniqueName: \"kubernetes.io/projected/f72cd798-9991-4f53-84ba-6a391217a278-kube-api-access-qthvq\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.313172 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-config\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.313333 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-scripts\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.313372 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.313387 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.313403 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f72cd798-9991-4f53-84ba-6a391217a278-logs\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.313418 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-dns-svc\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.314206 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f72cd798-9991-4f53-84ba-6a391217a278-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.314246 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data-custom\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.314157 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-sb\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.314159 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-config\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.313944 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-nb\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.314526 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-dns-svc\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.333897 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjs5d\" (UniqueName: \"kubernetes.io/projected/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-kube-api-access-gjs5d\") pod \"dnsmasq-dns-7b4c94945c-2ck29\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.380067 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.415740 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.415781 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.415802 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f72cd798-9991-4f53-84ba-6a391217a278-logs\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.415823 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f72cd798-9991-4f53-84ba-6a391217a278-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.415864 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data-custom\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.415942 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qthvq\" (UniqueName: \"kubernetes.io/projected/f72cd798-9991-4f53-84ba-6a391217a278-kube-api-access-qthvq\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.416108 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-scripts\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.418494 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f72cd798-9991-4f53-84ba-6a391217a278-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.418933 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f72cd798-9991-4f53-84ba-6a391217a278-logs\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.422433 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-scripts\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.422868 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data-custom\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.424860 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.425802 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.439655 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qthvq\" (UniqueName: \"kubernetes.io/projected/f72cd798-9991-4f53-84ba-6a391217a278-kube-api-access-qthvq\") pod \"cinder-api-0\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.522296 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.677309 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b4c94945c-2ck29"] Nov 24 10:23:46 crc kubenswrapper[4944]: I1124 10:23:46.748176 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" event={"ID":"2d60e5a3-02ae-4c67-8e12-2b0a988cc889","Type":"ContainerStarted","Data":"724132bf0945d8c7310814cfdfb0930aea4967e6b934b822ce5cd27f3db4b63a"} Nov 24 10:23:47 crc kubenswrapper[4944]: I1124 10:23:47.122215 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 10:23:47 crc kubenswrapper[4944]: W1124 10:23:47.124156 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf72cd798_9991_4f53_84ba_6a391217a278.slice/crio-e8f9a014164659ba22f625f20fdbeb0d3509985cc8f4698fcbb474b6d537f5fc WatchSource:0}: Error finding container e8f9a014164659ba22f625f20fdbeb0d3509985cc8f4698fcbb474b6d537f5fc: Status 404 returned error can't find the container with id e8f9a014164659ba22f625f20fdbeb0d3509985cc8f4698fcbb474b6d537f5fc Nov 24 10:23:47 crc kubenswrapper[4944]: I1124 10:23:47.756018 4944 generic.go:334] "Generic (PLEG): container finished" podID="2d60e5a3-02ae-4c67-8e12-2b0a988cc889" containerID="5a34dfb2847fe4bd8fc30d2886ce66b670b9581eaf95d0ce38a9221962e3f02d" exitCode=0 Nov 24 10:23:47 crc kubenswrapper[4944]: I1124 10:23:47.756087 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" event={"ID":"2d60e5a3-02ae-4c67-8e12-2b0a988cc889","Type":"ContainerDied","Data":"5a34dfb2847fe4bd8fc30d2886ce66b670b9581eaf95d0ce38a9221962e3f02d"} Nov 24 10:23:47 crc kubenswrapper[4944]: I1124 10:23:47.766929 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f72cd798-9991-4f53-84ba-6a391217a278","Type":"ContainerStarted","Data":"2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b"} Nov 24 10:23:47 crc kubenswrapper[4944]: I1124 10:23:47.766982 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f72cd798-9991-4f53-84ba-6a391217a278","Type":"ContainerStarted","Data":"e8f9a014164659ba22f625f20fdbeb0d3509985cc8f4698fcbb474b6d537f5fc"} Nov 24 10:23:48 crc kubenswrapper[4944]: I1124 10:23:48.776579 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" event={"ID":"2d60e5a3-02ae-4c67-8e12-2b0a988cc889","Type":"ContainerStarted","Data":"cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b"} Nov 24 10:23:48 crc kubenswrapper[4944]: I1124 10:23:48.777694 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:48 crc kubenswrapper[4944]: I1124 10:23:48.786939 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f72cd798-9991-4f53-84ba-6a391217a278","Type":"ContainerStarted","Data":"b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df"} Nov 24 10:23:48 crc kubenswrapper[4944]: I1124 10:23:48.788094 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 10:23:48 crc kubenswrapper[4944]: I1124 10:23:48.802016 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" podStartSLOduration=2.801996178 podStartE2EDuration="2.801996178s" podCreationTimestamp="2025-11-24 10:23:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:23:48.79736773 +0000 UTC m=+5489.331808192" watchObservedRunningTime="2025-11-24 10:23:48.801996178 +0000 UTC m=+5489.336436640" Nov 24 10:23:48 crc kubenswrapper[4944]: I1124 10:23:48.829706 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.82968525 podStartE2EDuration="2.82968525s" podCreationTimestamp="2025-11-24 10:23:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:23:48.819462805 +0000 UTC m=+5489.353903267" watchObservedRunningTime="2025-11-24 10:23:48.82968525 +0000 UTC m=+5489.364125712" Nov 24 10:23:53 crc kubenswrapper[4944]: I1124 10:23:53.549345 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:23:53 crc kubenswrapper[4944]: I1124 10:23:53.550131 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:23:53 crc kubenswrapper[4944]: I1124 10:23:53.550198 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 10:23:53 crc kubenswrapper[4944]: I1124 10:23:53.551360 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3b50540cdc5646c23244da47ca4ab4ee1e53c396cc0d5f9070264c91666bc8f2"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 10:23:53 crc kubenswrapper[4944]: I1124 10:23:53.551432 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://3b50540cdc5646c23244da47ca4ab4ee1e53c396cc0d5f9070264c91666bc8f2" gracePeriod=600 Nov 24 10:23:53 crc kubenswrapper[4944]: I1124 10:23:53.828826 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="3b50540cdc5646c23244da47ca4ab4ee1e53c396cc0d5f9070264c91666bc8f2" exitCode=0 Nov 24 10:23:53 crc kubenswrapper[4944]: I1124 10:23:53.828902 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"3b50540cdc5646c23244da47ca4ab4ee1e53c396cc0d5f9070264c91666bc8f2"} Nov 24 10:23:53 crc kubenswrapper[4944]: I1124 10:23:53.829119 4944 scope.go:117] "RemoveContainer" containerID="9c40a3d3e16cbf27ea9520f5d618ce4fd8655604883862f70ee96706277566bf" Nov 24 10:23:54 crc kubenswrapper[4944]: I1124 10:23:54.838954 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754"} Nov 24 10:23:56 crc kubenswrapper[4944]: I1124 10:23:56.382287 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:23:56 crc kubenswrapper[4944]: I1124 10:23:56.443223 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-574d65f5bf-8sw2n"] Nov 24 10:23:56 crc kubenswrapper[4944]: I1124 10:23:56.443439 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" podUID="a797b85c-0a33-4a54-a3cd-245b74bacba4" containerName="dnsmasq-dns" containerID="cri-o://68556ce24ecb94075a6493bb376a6960468563d64756480c2369c11ba32382b9" gracePeriod=10 Nov 24 10:23:56 crc kubenswrapper[4944]: I1124 10:23:56.864532 4944 generic.go:334] "Generic (PLEG): container finished" podID="a797b85c-0a33-4a54-a3cd-245b74bacba4" containerID="68556ce24ecb94075a6493bb376a6960468563d64756480c2369c11ba32382b9" exitCode=0 Nov 24 10:23:56 crc kubenswrapper[4944]: I1124 10:23:56.864847 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" event={"ID":"a797b85c-0a33-4a54-a3cd-245b74bacba4","Type":"ContainerDied","Data":"68556ce24ecb94075a6493bb376a6960468563d64756480c2369c11ba32382b9"} Nov 24 10:23:56 crc kubenswrapper[4944]: I1124 10:23:56.999014 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.009643 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-config\") pod \"a797b85c-0a33-4a54-a3cd-245b74bacba4\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.009718 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-dns-svc\") pod \"a797b85c-0a33-4a54-a3cd-245b74bacba4\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.009776 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-nb\") pod \"a797b85c-0a33-4a54-a3cd-245b74bacba4\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.084895 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a797b85c-0a33-4a54-a3cd-245b74bacba4" (UID: "a797b85c-0a33-4a54-a3cd-245b74bacba4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.094318 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a797b85c-0a33-4a54-a3cd-245b74bacba4" (UID: "a797b85c-0a33-4a54-a3cd-245b74bacba4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.113673 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-config" (OuterVolumeSpecName: "config") pod "a797b85c-0a33-4a54-a3cd-245b74bacba4" (UID: "a797b85c-0a33-4a54-a3cd-245b74bacba4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.113993 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-sb\") pod \"a797b85c-0a33-4a54-a3cd-245b74bacba4\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.114086 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckslw\" (UniqueName: \"kubernetes.io/projected/a797b85c-0a33-4a54-a3cd-245b74bacba4-kube-api-access-ckslw\") pod \"a797b85c-0a33-4a54-a3cd-245b74bacba4\" (UID: \"a797b85c-0a33-4a54-a3cd-245b74bacba4\") " Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.115777 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.115802 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.115811 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.135510 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a797b85c-0a33-4a54-a3cd-245b74bacba4-kube-api-access-ckslw" (OuterVolumeSpecName: "kube-api-access-ckslw") pod "a797b85c-0a33-4a54-a3cd-245b74bacba4" (UID: "a797b85c-0a33-4a54-a3cd-245b74bacba4"). InnerVolumeSpecName "kube-api-access-ckslw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.162868 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a797b85c-0a33-4a54-a3cd-245b74bacba4" (UID: "a797b85c-0a33-4a54-a3cd-245b74bacba4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.217974 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a797b85c-0a33-4a54-a3cd-245b74bacba4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.218024 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckslw\" (UniqueName: \"kubernetes.io/projected/a797b85c-0a33-4a54-a3cd-245b74bacba4-kube-api-access-ckslw\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.876534 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" event={"ID":"a797b85c-0a33-4a54-a3cd-245b74bacba4","Type":"ContainerDied","Data":"e146f726d62b3c8ec6ae2b0f2ba798b8326b5eaa927dbd6e1df07f3fb8f096aa"} Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.876605 4944 scope.go:117] "RemoveContainer" containerID="68556ce24ecb94075a6493bb376a6960468563d64756480c2369c11ba32382b9" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.876715 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-574d65f5bf-8sw2n" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.904844 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.905081 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="497cebe3-1e6d-4294-bef1-45453ace62f0" containerName="nova-scheduler-scheduler" containerID="cri-o://3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f" gracePeriod=30 Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.910779 4944 scope.go:117] "RemoveContainer" containerID="af3f54a378fe3878876ef46b3ad3b36dc93cd8e9db0901ca8eca74b0f84e213d" Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.927238 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.927482 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-log" containerID="cri-o://a79308575fa58b1f06e8a3c2da13f2e44ce9fa9b5031db96d89f6ca9517c741f" gracePeriod=30 Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.927545 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-metadata" containerID="cri-o://fbb488a6adfa0c34cce9196736c83a62353ec6209398f2fea648fb462c89a43b" gracePeriod=30 Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.939028 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.939388 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="a46345a8-68da-4344-84cc-f535d23c70c7" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://328b844e22c2e9064ca537cdd1059002c631f209faf697bfa41a5fd590d0cac4" gracePeriod=30 Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.947753 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.948164 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="5ddeb86e-6b6a-436f-929c-f53dca84ff3a" containerName="nova-cell1-conductor-conductor" containerID="cri-o://998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21" gracePeriod=30 Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.960174 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-574d65f5bf-8sw2n"] Nov 24 10:23:57 crc kubenswrapper[4944]: I1124 10:23:57.966812 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-574d65f5bf-8sw2n"] Nov 24 10:23:58 crc kubenswrapper[4944]: I1124 10:23:58.004814 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:23:58 crc kubenswrapper[4944]: I1124 10:23:58.005822 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-api" containerID="cri-o://58bc6140c220ada0743bf577e4ee73036b60a7bf502be440a2e8a7539fa4abf5" gracePeriod=30 Nov 24 10:23:58 crc kubenswrapper[4944]: I1124 10:23:58.006200 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-log" containerID="cri-o://0d5e1d6812a5dcca4341f23087431fa85f7b0889b653a4012082927656c833b6" gracePeriod=30 Nov 24 10:23:58 crc kubenswrapper[4944]: I1124 10:23:58.289496 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a797b85c-0a33-4a54-a3cd-245b74bacba4" path="/var/lib/kubelet/pods/a797b85c-0a33-4a54-a3cd-245b74bacba4/volumes" Nov 24 10:23:58 crc kubenswrapper[4944]: I1124 10:23:58.875570 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 10:23:58 crc kubenswrapper[4944]: I1124 10:23:58.905328 4944 generic.go:334] "Generic (PLEG): container finished" podID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerID="0d5e1d6812a5dcca4341f23087431fa85f7b0889b653a4012082927656c833b6" exitCode=143 Nov 24 10:23:58 crc kubenswrapper[4944]: I1124 10:23:58.905473 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"401b7639-8818-40f5-9c89-cbeacfcf465a","Type":"ContainerDied","Data":"0d5e1d6812a5dcca4341f23087431fa85f7b0889b653a4012082927656c833b6"} Nov 24 10:23:58 crc kubenswrapper[4944]: I1124 10:23:58.916846 4944 generic.go:334] "Generic (PLEG): container finished" podID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerID="a79308575fa58b1f06e8a3c2da13f2e44ce9fa9b5031db96d89f6ca9517c741f" exitCode=143 Nov 24 10:23:58 crc kubenswrapper[4944]: I1124 10:23:58.916940 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b8d3337-080c-4f51-be4d-9a69f6a6e074","Type":"ContainerDied","Data":"a79308575fa58b1f06e8a3c2da13f2e44ce9fa9b5031db96d89f6ca9517c741f"} Nov 24 10:23:58 crc kubenswrapper[4944]: I1124 10:23:58.919271 4944 generic.go:334] "Generic (PLEG): container finished" podID="a46345a8-68da-4344-84cc-f535d23c70c7" containerID="328b844e22c2e9064ca537cdd1059002c631f209faf697bfa41a5fd590d0cac4" exitCode=0 Nov 24 10:23:58 crc kubenswrapper[4944]: I1124 10:23:58.919299 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a46345a8-68da-4344-84cc-f535d23c70c7","Type":"ContainerDied","Data":"328b844e22c2e9064ca537cdd1059002c631f209faf697bfa41a5fd590d0cac4"} Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.070477 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.177804 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-combined-ca-bundle\") pod \"a46345a8-68da-4344-84cc-f535d23c70c7\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.177873 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-config-data\") pod \"a46345a8-68da-4344-84cc-f535d23c70c7\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.177975 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-964gp\" (UniqueName: \"kubernetes.io/projected/a46345a8-68da-4344-84cc-f535d23c70c7-kube-api-access-964gp\") pod \"a46345a8-68da-4344-84cc-f535d23c70c7\" (UID: \"a46345a8-68da-4344-84cc-f535d23c70c7\") " Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.185361 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a46345a8-68da-4344-84cc-f535d23c70c7-kube-api-access-964gp" (OuterVolumeSpecName: "kube-api-access-964gp") pod "a46345a8-68da-4344-84cc-f535d23c70c7" (UID: "a46345a8-68da-4344-84cc-f535d23c70c7"). InnerVolumeSpecName "kube-api-access-964gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.239082 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-config-data" (OuterVolumeSpecName: "config-data") pod "a46345a8-68da-4344-84cc-f535d23c70c7" (UID: "a46345a8-68da-4344-84cc-f535d23c70c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.240899 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a46345a8-68da-4344-84cc-f535d23c70c7" (UID: "a46345a8-68da-4344-84cc-f535d23c70c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.279378 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.279408 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a46345a8-68da-4344-84cc-f535d23c70c7-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.279417 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-964gp\" (UniqueName: \"kubernetes.io/projected/a46345a8-68da-4344-84cc-f535d23c70c7-kube-api-access-964gp\") on node \"crc\" DevicePath \"\"" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.928450 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a46345a8-68da-4344-84cc-f535d23c70c7","Type":"ContainerDied","Data":"7386522d0c1ac192064c05dbd01d6bbd7dcd54dadd417be615cfdcbfedc0ba3d"} Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.928497 4944 scope.go:117] "RemoveContainer" containerID="328b844e22c2e9064ca537cdd1059002c631f209faf697bfa41a5fd590d0cac4" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.928603 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.969224 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.980836 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.993866 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 10:23:59 crc kubenswrapper[4944]: E1124 10:23:59.994786 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a797b85c-0a33-4a54-a3cd-245b74bacba4" containerName="dnsmasq-dns" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.994808 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a797b85c-0a33-4a54-a3cd-245b74bacba4" containerName="dnsmasq-dns" Nov 24 10:23:59 crc kubenswrapper[4944]: E1124 10:23:59.994830 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a46345a8-68da-4344-84cc-f535d23c70c7" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.994838 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a46345a8-68da-4344-84cc-f535d23c70c7" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 10:23:59 crc kubenswrapper[4944]: E1124 10:23:59.994873 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a797b85c-0a33-4a54-a3cd-245b74bacba4" containerName="init" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.994881 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="a797b85c-0a33-4a54-a3cd-245b74bacba4" containerName="init" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.995130 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a797b85c-0a33-4a54-a3cd-245b74bacba4" containerName="dnsmasq-dns" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.995168 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="a46345a8-68da-4344-84cc-f535d23c70c7" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.996329 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:23:59 crc kubenswrapper[4944]: I1124 10:23:59.999127 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.004089 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.195028 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74lw7\" (UniqueName: \"kubernetes.io/projected/3c726090-4f6f-40db-a2c2-6217339f122d-kube-api-access-74lw7\") pod \"nova-cell1-novncproxy-0\" (UID: \"3c726090-4f6f-40db-a2c2-6217339f122d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.195089 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c726090-4f6f-40db-a2c2-6217339f122d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3c726090-4f6f-40db-a2c2-6217339f122d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.195269 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c726090-4f6f-40db-a2c2-6217339f122d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3c726090-4f6f-40db-a2c2-6217339f122d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.290539 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a46345a8-68da-4344-84cc-f535d23c70c7" path="/var/lib/kubelet/pods/a46345a8-68da-4344-84cc-f535d23c70c7/volumes" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.296551 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74lw7\" (UniqueName: \"kubernetes.io/projected/3c726090-4f6f-40db-a2c2-6217339f122d-kube-api-access-74lw7\") pod \"nova-cell1-novncproxy-0\" (UID: \"3c726090-4f6f-40db-a2c2-6217339f122d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.296607 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c726090-4f6f-40db-a2c2-6217339f122d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3c726090-4f6f-40db-a2c2-6217339f122d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.296674 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c726090-4f6f-40db-a2c2-6217339f122d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3c726090-4f6f-40db-a2c2-6217339f122d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.311039 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c726090-4f6f-40db-a2c2-6217339f122d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3c726090-4f6f-40db-a2c2-6217339f122d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.311058 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c726090-4f6f-40db-a2c2-6217339f122d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3c726090-4f6f-40db-a2c2-6217339f122d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.326884 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74lw7\" (UniqueName: \"kubernetes.io/projected/3c726090-4f6f-40db-a2c2-6217339f122d-kube-api-access-74lw7\") pod \"nova-cell1-novncproxy-0\" (UID: \"3c726090-4f6f-40db-a2c2-6217339f122d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.612575 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.706720 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.805108 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-combined-ca-bundle\") pod \"497cebe3-1e6d-4294-bef1-45453ace62f0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.805180 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4dv9\" (UniqueName: \"kubernetes.io/projected/497cebe3-1e6d-4294-bef1-45453ace62f0-kube-api-access-t4dv9\") pod \"497cebe3-1e6d-4294-bef1-45453ace62f0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.805237 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-config-data\") pod \"497cebe3-1e6d-4294-bef1-45453ace62f0\" (UID: \"497cebe3-1e6d-4294-bef1-45453ace62f0\") " Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.810728 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/497cebe3-1e6d-4294-bef1-45453ace62f0-kube-api-access-t4dv9" (OuterVolumeSpecName: "kube-api-access-t4dv9") pod "497cebe3-1e6d-4294-bef1-45453ace62f0" (UID: "497cebe3-1e6d-4294-bef1-45453ace62f0"). InnerVolumeSpecName "kube-api-access-t4dv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.858361 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "497cebe3-1e6d-4294-bef1-45453ace62f0" (UID: "497cebe3-1e6d-4294-bef1-45453ace62f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.861828 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-config-data" (OuterVolumeSpecName: "config-data") pod "497cebe3-1e6d-4294-bef1-45453ace62f0" (UID: "497cebe3-1e6d-4294-bef1-45453ace62f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.907306 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.907336 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4dv9\" (UniqueName: \"kubernetes.io/projected/497cebe3-1e6d-4294-bef1-45453ace62f0-kube-api-access-t4dv9\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.907350 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/497cebe3-1e6d-4294-bef1-45453ace62f0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.940545 4944 generic.go:334] "Generic (PLEG): container finished" podID="497cebe3-1e6d-4294-bef1-45453ace62f0" containerID="3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f" exitCode=0 Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.940633 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.940636 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"497cebe3-1e6d-4294-bef1-45453ace62f0","Type":"ContainerDied","Data":"3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f"} Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.940689 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"497cebe3-1e6d-4294-bef1-45453ace62f0","Type":"ContainerDied","Data":"d58d58a7748545c88ed14f5e7f8d5204d62af14bc6c13975e1407a289f5bdff0"} Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.940707 4944 scope.go:117] "RemoveContainer" containerID="3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.976415 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.978993 4944 scope.go:117] "RemoveContainer" containerID="3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f" Nov 24 10:24:00 crc kubenswrapper[4944]: E1124 10:24:00.979606 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f\": container with ID starting with 3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f not found: ID does not exist" containerID="3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.979658 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f"} err="failed to get container status \"3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f\": rpc error: code = NotFound desc = could not find container \"3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f\": container with ID starting with 3500f8ecc3866c16faa2cc60749b6ea5dbd537556eab17d3cb0b1b0ae528011f not found: ID does not exist" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.989880 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.996269 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:24:00 crc kubenswrapper[4944]: E1124 10:24:00.996964 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="497cebe3-1e6d-4294-bef1-45453ace62f0" containerName="nova-scheduler-scheduler" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.997070 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="497cebe3-1e6d-4294-bef1-45453ace62f0" containerName="nova-scheduler-scheduler" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.997339 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="497cebe3-1e6d-4294-bef1-45453ace62f0" containerName="nova-scheduler-scheduler" Nov 24 10:24:00 crc kubenswrapper[4944]: I1124 10:24:00.998042 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.000014 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.007635 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.096257 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.110634 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-config-data\") pod \"nova-scheduler-0\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " pod="openstack/nova-scheduler-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.110679 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " pod="openstack/nova-scheduler-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.110696 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85r4p\" (UniqueName: \"kubernetes.io/projected/f028a568-aeb6-435b-81f5-d5014b3609c3-kube-api-access-85r4p\") pod \"nova-scheduler-0\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " pod="openstack/nova-scheduler-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.212521 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-config-data\") pod \"nova-scheduler-0\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " pod="openstack/nova-scheduler-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.212574 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " pod="openstack/nova-scheduler-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.213030 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85r4p\" (UniqueName: \"kubernetes.io/projected/f028a568-aeb6-435b-81f5-d5014b3609c3-kube-api-access-85r4p\") pod \"nova-scheduler-0\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " pod="openstack/nova-scheduler-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.215528 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-config-data\") pod \"nova-scheduler-0\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " pod="openstack/nova-scheduler-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.218187 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " pod="openstack/nova-scheduler-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.228154 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85r4p\" (UniqueName: \"kubernetes.io/projected/f028a568-aeb6-435b-81f5-d5014b3609c3-kube-api-access-85r4p\") pod \"nova-scheduler-0\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " pod="openstack/nova-scheduler-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.359965 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.451815 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.519667 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-config-data\") pod \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.519726 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzv4r\" (UniqueName: \"kubernetes.io/projected/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-kube-api-access-dzv4r\") pod \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.519824 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-combined-ca-bundle\") pod \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\" (UID: \"5ddeb86e-6b6a-436f-929c-f53dca84ff3a\") " Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.531766 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-kube-api-access-dzv4r" (OuterVolumeSpecName: "kube-api-access-dzv4r") pod "5ddeb86e-6b6a-436f-929c-f53dca84ff3a" (UID: "5ddeb86e-6b6a-436f-929c-f53dca84ff3a"). InnerVolumeSpecName "kube-api-access-dzv4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.547979 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ddeb86e-6b6a-436f-929c-f53dca84ff3a" (UID: "5ddeb86e-6b6a-436f-929c-f53dca84ff3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.555789 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-config-data" (OuterVolumeSpecName: "config-data") pod "5ddeb86e-6b6a-436f-929c-f53dca84ff3a" (UID: "5ddeb86e-6b6a-436f-929c-f53dca84ff3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.622664 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.622982 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.622992 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzv4r\" (UniqueName: \"kubernetes.io/projected/5ddeb86e-6b6a-436f-929c-f53dca84ff3a-kube-api-access-dzv4r\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.624750 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.624957 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="f696e4c8-bb0c-410e-af8d-351a42bee463" containerName="nova-cell0-conductor-conductor" containerID="cri-o://cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718" gracePeriod=30 Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.715863 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.72:8774/\": dial tcp 10.217.1.72:8774: connect: connection refused" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.715882 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.72:8774/\": dial tcp 10.217.1.72:8774: connect: connection refused" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.966203 4944 generic.go:334] "Generic (PLEG): container finished" podID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerID="58bc6140c220ada0743bf577e4ee73036b60a7bf502be440a2e8a7539fa4abf5" exitCode=0 Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.966285 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"401b7639-8818-40f5-9c89-cbeacfcf465a","Type":"ContainerDied","Data":"58bc6140c220ada0743bf577e4ee73036b60a7bf502be440a2e8a7539fa4abf5"} Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.978979 4944 generic.go:334] "Generic (PLEG): container finished" podID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerID="fbb488a6adfa0c34cce9196736c83a62353ec6209398f2fea648fb462c89a43b" exitCode=0 Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.979035 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b8d3337-080c-4f51-be4d-9a69f6a6e074","Type":"ContainerDied","Data":"fbb488a6adfa0c34cce9196736c83a62353ec6209398f2fea648fb462c89a43b"} Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.979073 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b8d3337-080c-4f51-be4d-9a69f6a6e074","Type":"ContainerDied","Data":"a526a1fdf2b77de9aa2e9601dead235ea4aecb9d51497a4ab42ef52dec1453fe"} Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.979085 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a526a1fdf2b77de9aa2e9601dead235ea4aecb9d51497a4ab42ef52dec1453fe" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.982703 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3c726090-4f6f-40db-a2c2-6217339f122d","Type":"ContainerStarted","Data":"ea9c1960080c67b944eb7b32a45c16360f75c79d59c597093ff83a5efced6975"} Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.982754 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3c726090-4f6f-40db-a2c2-6217339f122d","Type":"ContainerStarted","Data":"9c7cb4cc61887772972d8810087ee35a9cd4994e793af18c91edd04dc9ee8fe8"} Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.987088 4944 generic.go:334] "Generic (PLEG): container finished" podID="5ddeb86e-6b6a-436f-929c-f53dca84ff3a" containerID="998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21" exitCode=0 Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.987153 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5ddeb86e-6b6a-436f-929c-f53dca84ff3a","Type":"ContainerDied","Data":"998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21"} Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.987179 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5ddeb86e-6b6a-436f-929c-f53dca84ff3a","Type":"ContainerDied","Data":"5446698ae40f45dc173acca166e8dfe4c979116e9f277f406461aa336975cce6"} Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.987202 4944 scope.go:117] "RemoveContainer" containerID="998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21" Nov 24 10:24:01 crc kubenswrapper[4944]: I1124 10:24:01.987312 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.008897 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.008810996 podStartE2EDuration="3.008810996s" podCreationTimestamp="2025-11-24 10:23:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:24:02.002806645 +0000 UTC m=+5502.537247107" watchObservedRunningTime="2025-11-24 10:24:02.008810996 +0000 UTC m=+5502.543251458" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.028411 4944 scope.go:117] "RemoveContainer" containerID="998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21" Nov 24 10:24:02 crc kubenswrapper[4944]: E1124 10:24:02.031637 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21\": container with ID starting with 998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21 not found: ID does not exist" containerID="998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.031694 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21"} err="failed to get container status \"998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21\": rpc error: code = NotFound desc = could not find container \"998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21\": container with ID starting with 998caa0e59a3c63a6799bca791af649c3bd5e08aaf77f3ae08e4f755a98aad21 not found: ID does not exist" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.053778 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.093665 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.106819 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.108662 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.113702 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.139929 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 10:24:02 crc kubenswrapper[4944]: E1124 10:24:02.140420 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-log" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.140453 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-log" Nov 24 10:24:02 crc kubenswrapper[4944]: E1124 10:24:02.140481 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-log" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.140488 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-log" Nov 24 10:24:02 crc kubenswrapper[4944]: E1124 10:24:02.140501 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-metadata" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.140507 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-metadata" Nov 24 10:24:02 crc kubenswrapper[4944]: E1124 10:24:02.140537 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ddeb86e-6b6a-436f-929c-f53dca84ff3a" containerName="nova-cell1-conductor-conductor" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.140544 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ddeb86e-6b6a-436f-929c-f53dca84ff3a" containerName="nova-cell1-conductor-conductor" Nov 24 10:24:02 crc kubenswrapper[4944]: E1124 10:24:02.140560 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-api" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.140565 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-api" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.140792 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-metadata" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.140802 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-log" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.140812 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-log" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.140822 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ddeb86e-6b6a-436f-929c-f53dca84ff3a" containerName="nova-cell1-conductor-conductor" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.140847 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" containerName="nova-api-api" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.141634 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.143848 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.170159 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.235866 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-combined-ca-bundle\") pod \"401b7639-8818-40f5-9c89-cbeacfcf465a\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.235941 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8dz9\" (UniqueName: \"kubernetes.io/projected/3b8d3337-080c-4f51-be4d-9a69f6a6e074-kube-api-access-p8dz9\") pod \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.236086 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b8d3337-080c-4f51-be4d-9a69f6a6e074-logs\") pod \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.236135 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/401b7639-8818-40f5-9c89-cbeacfcf465a-logs\") pod \"401b7639-8818-40f5-9c89-cbeacfcf465a\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.236206 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-combined-ca-bundle\") pod \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.236235 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-config-data\") pod \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\" (UID: \"3b8d3337-080c-4f51-be4d-9a69f6a6e074\") " Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.236295 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf6vs\" (UniqueName: \"kubernetes.io/projected/401b7639-8818-40f5-9c89-cbeacfcf465a-kube-api-access-hf6vs\") pod \"401b7639-8818-40f5-9c89-cbeacfcf465a\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.236342 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-config-data\") pod \"401b7639-8818-40f5-9c89-cbeacfcf465a\" (UID: \"401b7639-8818-40f5-9c89-cbeacfcf465a\") " Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.236646 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.236724 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw2hw\" (UniqueName: \"kubernetes.io/projected/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-kube-api-access-qw2hw\") pod \"nova-cell1-conductor-0\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.236817 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.237705 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b8d3337-080c-4f51-be4d-9a69f6a6e074-logs" (OuterVolumeSpecName: "logs") pod "3b8d3337-080c-4f51-be4d-9a69f6a6e074" (UID: "3b8d3337-080c-4f51-be4d-9a69f6a6e074"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.237916 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/401b7639-8818-40f5-9c89-cbeacfcf465a-logs" (OuterVolumeSpecName: "logs") pod "401b7639-8818-40f5-9c89-cbeacfcf465a" (UID: "401b7639-8818-40f5-9c89-cbeacfcf465a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.241726 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401b7639-8818-40f5-9c89-cbeacfcf465a-kube-api-access-hf6vs" (OuterVolumeSpecName: "kube-api-access-hf6vs") pod "401b7639-8818-40f5-9c89-cbeacfcf465a" (UID: "401b7639-8818-40f5-9c89-cbeacfcf465a"). InnerVolumeSpecName "kube-api-access-hf6vs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.243782 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8d3337-080c-4f51-be4d-9a69f6a6e074-kube-api-access-p8dz9" (OuterVolumeSpecName: "kube-api-access-p8dz9") pod "3b8d3337-080c-4f51-be4d-9a69f6a6e074" (UID: "3b8d3337-080c-4f51-be4d-9a69f6a6e074"). InnerVolumeSpecName "kube-api-access-p8dz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.286181 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b8d3337-080c-4f51-be4d-9a69f6a6e074" (UID: "3b8d3337-080c-4f51-be4d-9a69f6a6e074"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.296227 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="497cebe3-1e6d-4294-bef1-45453ace62f0" path="/var/lib/kubelet/pods/497cebe3-1e6d-4294-bef1-45453ace62f0/volumes" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.297494 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ddeb86e-6b6a-436f-929c-f53dca84ff3a" path="/var/lib/kubelet/pods/5ddeb86e-6b6a-436f-929c-f53dca84ff3a/volumes" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.297815 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-config-data" (OuterVolumeSpecName: "config-data") pod "401b7639-8818-40f5-9c89-cbeacfcf465a" (UID: "401b7639-8818-40f5-9c89-cbeacfcf465a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.301172 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "401b7639-8818-40f5-9c89-cbeacfcf465a" (UID: "401b7639-8818-40f5-9c89-cbeacfcf465a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.308786 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-config-data" (OuterVolumeSpecName: "config-data") pod "3b8d3337-080c-4f51-be4d-9a69f6a6e074" (UID: "3b8d3337-080c-4f51-be4d-9a69f6a6e074"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.338605 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.338728 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.338795 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw2hw\" (UniqueName: \"kubernetes.io/projected/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-kube-api-access-qw2hw\") pod \"nova-cell1-conductor-0\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.338912 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.338935 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8dz9\" (UniqueName: \"kubernetes.io/projected/3b8d3337-080c-4f51-be4d-9a69f6a6e074-kube-api-access-p8dz9\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.338948 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b8d3337-080c-4f51-be4d-9a69f6a6e074-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.338959 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/401b7639-8818-40f5-9c89-cbeacfcf465a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.338971 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.338981 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8d3337-080c-4f51-be4d-9a69f6a6e074-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.338992 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf6vs\" (UniqueName: \"kubernetes.io/projected/401b7639-8818-40f5-9c89-cbeacfcf465a-kube-api-access-hf6vs\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.339002 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/401b7639-8818-40f5-9c89-cbeacfcf465a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.343896 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.344115 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.359268 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw2hw\" (UniqueName: \"kubernetes.io/projected/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-kube-api-access-qw2hw\") pod \"nova-cell1-conductor-0\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.472743 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:02 crc kubenswrapper[4944]: I1124 10:24:02.918077 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 10:24:02 crc kubenswrapper[4944]: E1124 10:24:02.995915 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 10:24:02 crc kubenswrapper[4944]: E1124 10:24:02.999049 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 10:24:03 crc kubenswrapper[4944]: E1124 10:24:03.000573 4944 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 24 10:24:03 crc kubenswrapper[4944]: E1124 10:24:03.000664 4944 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="f696e4c8-bb0c-410e-af8d-351a42bee463" containerName="nova-cell0-conductor-conductor" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.010621 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e8551f1e-3692-4d2a-99b3-1d1015b35e9c","Type":"ContainerStarted","Data":"c2238a3452b6725552bdcd3e6ab24ca43238d800d732c1e6c9c7c16a575109ea"} Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.016168 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f028a568-aeb6-435b-81f5-d5014b3609c3","Type":"ContainerStarted","Data":"520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81"} Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.016210 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f028a568-aeb6-435b-81f5-d5014b3609c3","Type":"ContainerStarted","Data":"533851fee3ac440c0c311eb6b73c444e1655970aeb0d69e7e82c9dd7fcc498bd"} Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.033724 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.034904 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"401b7639-8818-40f5-9c89-cbeacfcf465a","Type":"ContainerDied","Data":"684b55fc4bfaf8641f206857eb3d5754a746e35b95b5ca3a2a8db846aec74a0f"} Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.034951 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.034967 4944 scope.go:117] "RemoveContainer" containerID="58bc6140c220ada0743bf577e4ee73036b60a7bf502be440a2e8a7539fa4abf5" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.099220 4944 scope.go:117] "RemoveContainer" containerID="0d5e1d6812a5dcca4341f23087431fa85f7b0889b653a4012082927656c833b6" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.143687 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.143673403 podStartE2EDuration="3.143673403s" podCreationTimestamp="2025-11-24 10:24:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:24:03.061836157 +0000 UTC m=+5503.596276629" watchObservedRunningTime="2025-11-24 10:24:03.143673403 +0000 UTC m=+5503.678113865" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.154487 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.179129 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.195126 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.197002 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.201193 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.207876 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.220107 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.231435 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.244918 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.246864 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.248303 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.250870 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.276209 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64282688-4ab2-436d-bc65-97c100066cd2-logs\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.276308 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.276376 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-config-data\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.276400 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnsnn\" (UniqueName: \"kubernetes.io/projected/64282688-4ab2-436d-bc65-97c100066cd2-kube-api-access-fnsnn\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.379997 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64282688-4ab2-436d-bc65-97c100066cd2-logs\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.380162 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g59cv\" (UniqueName: \"kubernetes.io/projected/880181b3-27ad-4f6c-9471-6016ea2d7e75-kube-api-access-g59cv\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.380200 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/880181b3-27ad-4f6c-9471-6016ea2d7e75-logs\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.380234 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.380268 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-config-data\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.380301 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnsnn\" (UniqueName: \"kubernetes.io/projected/64282688-4ab2-436d-bc65-97c100066cd2-kube-api-access-fnsnn\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.380318 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-config-data\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.380377 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.380802 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64282688-4ab2-436d-bc65-97c100066cd2-logs\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.397850 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-config-data\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.404974 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnsnn\" (UniqueName: \"kubernetes.io/projected/64282688-4ab2-436d-bc65-97c100066cd2-kube-api-access-fnsnn\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.406804 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.481868 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-config-data\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.481957 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.482009 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g59cv\" (UniqueName: \"kubernetes.io/projected/880181b3-27ad-4f6c-9471-6016ea2d7e75-kube-api-access-g59cv\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.482040 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/880181b3-27ad-4f6c-9471-6016ea2d7e75-logs\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.482475 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/880181b3-27ad-4f6c-9471-6016ea2d7e75-logs\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.486799 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.486856 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-config-data\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.501169 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g59cv\" (UniqueName: \"kubernetes.io/projected/880181b3-27ad-4f6c-9471-6016ea2d7e75-kube-api-access-g59cv\") pod \"nova-metadata-0\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " pod="openstack/nova-metadata-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.529565 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 10:24:03 crc kubenswrapper[4944]: I1124 10:24:03.577995 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 10:24:04 crc kubenswrapper[4944]: I1124 10:24:04.044998 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e8551f1e-3692-4d2a-99b3-1d1015b35e9c","Type":"ContainerStarted","Data":"9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb"} Nov 24 10:24:04 crc kubenswrapper[4944]: I1124 10:24:04.045398 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:04 crc kubenswrapper[4944]: I1124 10:24:04.046375 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 10:24:04 crc kubenswrapper[4944]: I1124 10:24:04.069636 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.069611096 podStartE2EDuration="2.069611096s" podCreationTimestamp="2025-11-24 10:24:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:24:04.056336814 +0000 UTC m=+5504.590777286" watchObservedRunningTime="2025-11-24 10:24:04.069611096 +0000 UTC m=+5504.604051558" Nov 24 10:24:04 crc kubenswrapper[4944]: I1124 10:24:04.138337 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 10:24:04 crc kubenswrapper[4944]: W1124 10:24:04.151745 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod880181b3_27ad_4f6c_9471_6016ea2d7e75.slice/crio-ddbf2b9d8f7861ead89a5dc5d30a4a698e01b118d6ea1855b27ded05026526bb WatchSource:0}: Error finding container ddbf2b9d8f7861ead89a5dc5d30a4a698e01b118d6ea1855b27ded05026526bb: Status 404 returned error can't find the container with id ddbf2b9d8f7861ead89a5dc5d30a4a698e01b118d6ea1855b27ded05026526bb Nov 24 10:24:04 crc kubenswrapper[4944]: I1124 10:24:04.295895 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" path="/var/lib/kubelet/pods/3b8d3337-080c-4f51-be4d-9a69f6a6e074/volumes" Nov 24 10:24:04 crc kubenswrapper[4944]: I1124 10:24:04.297038 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="401b7639-8818-40f5-9c89-cbeacfcf465a" path="/var/lib/kubelet/pods/401b7639-8818-40f5-9c89-cbeacfcf465a/volumes" Nov 24 10:24:05 crc kubenswrapper[4944]: I1124 10:24:05.058872 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64282688-4ab2-436d-bc65-97c100066cd2","Type":"ContainerStarted","Data":"1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8"} Nov 24 10:24:05 crc kubenswrapper[4944]: I1124 10:24:05.059242 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64282688-4ab2-436d-bc65-97c100066cd2","Type":"ContainerStarted","Data":"651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676"} Nov 24 10:24:05 crc kubenswrapper[4944]: I1124 10:24:05.059255 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64282688-4ab2-436d-bc65-97c100066cd2","Type":"ContainerStarted","Data":"9c9dca476d4e35e27c127d04c21a6059576314298d613acf33f14688da7c5550"} Nov 24 10:24:05 crc kubenswrapper[4944]: I1124 10:24:05.062158 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"880181b3-27ad-4f6c-9471-6016ea2d7e75","Type":"ContainerStarted","Data":"37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426"} Nov 24 10:24:05 crc kubenswrapper[4944]: I1124 10:24:05.062234 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"880181b3-27ad-4f6c-9471-6016ea2d7e75","Type":"ContainerStarted","Data":"1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6"} Nov 24 10:24:05 crc kubenswrapper[4944]: I1124 10:24:05.062244 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"880181b3-27ad-4f6c-9471-6016ea2d7e75","Type":"ContainerStarted","Data":"ddbf2b9d8f7861ead89a5dc5d30a4a698e01b118d6ea1855b27ded05026526bb"} Nov 24 10:24:05 crc kubenswrapper[4944]: I1124 10:24:05.081088 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.081040762 podStartE2EDuration="2.081040762s" podCreationTimestamp="2025-11-24 10:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:24:05.078117359 +0000 UTC m=+5505.612557821" watchObservedRunningTime="2025-11-24 10:24:05.081040762 +0000 UTC m=+5505.615481224" Nov 24 10:24:05 crc kubenswrapper[4944]: I1124 10:24:05.099998 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.099976315 podStartE2EDuration="2.099976315s" podCreationTimestamp="2025-11-24 10:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:24:05.092648162 +0000 UTC m=+5505.627088624" watchObservedRunningTime="2025-11-24 10:24:05.099976315 +0000 UTC m=+5505.634416777" Nov 24 10:24:05 crc kubenswrapper[4944]: I1124 10:24:05.613233 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.360791 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.729389 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.813794 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.73:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.813802 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3b8d3337-080c-4f51-be4d-9a69f6a6e074" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.73:8775/\": dial tcp 10.217.1.73:8775: i/o timeout (Client.Timeout exceeded while awaiting headers)" Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.844030 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-config-data\") pod \"f696e4c8-bb0c-410e-af8d-351a42bee463\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.844139 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzqrl\" (UniqueName: \"kubernetes.io/projected/f696e4c8-bb0c-410e-af8d-351a42bee463-kube-api-access-hzqrl\") pod \"f696e4c8-bb0c-410e-af8d-351a42bee463\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.844216 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-combined-ca-bundle\") pod \"f696e4c8-bb0c-410e-af8d-351a42bee463\" (UID: \"f696e4c8-bb0c-410e-af8d-351a42bee463\") " Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.857189 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f696e4c8-bb0c-410e-af8d-351a42bee463-kube-api-access-hzqrl" (OuterVolumeSpecName: "kube-api-access-hzqrl") pod "f696e4c8-bb0c-410e-af8d-351a42bee463" (UID: "f696e4c8-bb0c-410e-af8d-351a42bee463"). InnerVolumeSpecName "kube-api-access-hzqrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.870363 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-config-data" (OuterVolumeSpecName: "config-data") pod "f696e4c8-bb0c-410e-af8d-351a42bee463" (UID: "f696e4c8-bb0c-410e-af8d-351a42bee463"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.872367 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f696e4c8-bb0c-410e-af8d-351a42bee463" (UID: "f696e4c8-bb0c-410e-af8d-351a42bee463"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.946755 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.946796 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzqrl\" (UniqueName: \"kubernetes.io/projected/f696e4c8-bb0c-410e-af8d-351a42bee463-kube-api-access-hzqrl\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:06 crc kubenswrapper[4944]: I1124 10:24:06.946810 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f696e4c8-bb0c-410e-af8d-351a42bee463-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.081654 4944 generic.go:334] "Generic (PLEG): container finished" podID="f696e4c8-bb0c-410e-af8d-351a42bee463" containerID="cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718" exitCode=0 Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.081716 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f696e4c8-bb0c-410e-af8d-351a42bee463","Type":"ContainerDied","Data":"cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718"} Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.081748 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.081773 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f696e4c8-bb0c-410e-af8d-351a42bee463","Type":"ContainerDied","Data":"0a293a3b644827d3823455ce0def1b3207837c25a0ff7c7b9b4bd8007dbfc7a2"} Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.081793 4944 scope.go:117] "RemoveContainer" containerID="cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.106557 4944 scope.go:117] "RemoveContainer" containerID="cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718" Nov 24 10:24:07 crc kubenswrapper[4944]: E1124 10:24:07.107066 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718\": container with ID starting with cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718 not found: ID does not exist" containerID="cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.107100 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718"} err="failed to get container status \"cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718\": rpc error: code = NotFound desc = could not find container \"cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718\": container with ID starting with cb11d94656073d49c70bed48850c6ca40a2f2eb05c1a0f690fd11bdc7ebe9718 not found: ID does not exist" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.125046 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.140814 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.146562 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 10:24:07 crc kubenswrapper[4944]: E1124 10:24:07.147089 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f696e4c8-bb0c-410e-af8d-351a42bee463" containerName="nova-cell0-conductor-conductor" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.147110 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f696e4c8-bb0c-410e-af8d-351a42bee463" containerName="nova-cell0-conductor-conductor" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.147338 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f696e4c8-bb0c-410e-af8d-351a42bee463" containerName="nova-cell0-conductor-conductor" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.148201 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.152318 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.160621 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.251711 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.251826 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7n99\" (UniqueName: \"kubernetes.io/projected/f7a89164-b59e-4690-b0e5-3482f54a769c-kube-api-access-c7n99\") pod \"nova-cell0-conductor-0\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.251858 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.353417 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.354335 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7n99\" (UniqueName: \"kubernetes.io/projected/f7a89164-b59e-4690-b0e5-3482f54a769c-kube-api-access-c7n99\") pod \"nova-cell0-conductor-0\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.354406 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.358343 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.359482 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.369123 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7n99\" (UniqueName: \"kubernetes.io/projected/f7a89164-b59e-4690-b0e5-3482f54a769c-kube-api-access-c7n99\") pod \"nova-cell0-conductor-0\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.479374 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:07 crc kubenswrapper[4944]: I1124 10:24:07.876921 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 10:24:08 crc kubenswrapper[4944]: I1124 10:24:08.092877 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f7a89164-b59e-4690-b0e5-3482f54a769c","Type":"ContainerStarted","Data":"9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2"} Nov 24 10:24:08 crc kubenswrapper[4944]: I1124 10:24:08.092926 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f7a89164-b59e-4690-b0e5-3482f54a769c","Type":"ContainerStarted","Data":"1fe0ffbfd7d4e5cb3acedd720e2ea41d0dcac778901253bd7120db4841fd78ad"} Nov 24 10:24:08 crc kubenswrapper[4944]: I1124 10:24:08.093329 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:08 crc kubenswrapper[4944]: I1124 10:24:08.111990 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.111968601 podStartE2EDuration="1.111968601s" podCreationTimestamp="2025-11-24 10:24:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:24:08.11096217 +0000 UTC m=+5508.645402642" watchObservedRunningTime="2025-11-24 10:24:08.111968601 +0000 UTC m=+5508.646409063" Nov 24 10:24:08 crc kubenswrapper[4944]: I1124 10:24:08.287693 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f696e4c8-bb0c-410e-af8d-351a42bee463" path="/var/lib/kubelet/pods/f696e4c8-bb0c-410e-af8d-351a42bee463/volumes" Nov 24 10:24:08 crc kubenswrapper[4944]: I1124 10:24:08.578085 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 10:24:08 crc kubenswrapper[4944]: I1124 10:24:08.578145 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 10:24:10 crc kubenswrapper[4944]: I1124 10:24:10.613382 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:10 crc kubenswrapper[4944]: I1124 10:24:10.626351 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:11 crc kubenswrapper[4944]: I1124 10:24:11.140143 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 10:24:11 crc kubenswrapper[4944]: I1124 10:24:11.360766 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 10:24:11 crc kubenswrapper[4944]: I1124 10:24:11.392100 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 10:24:12 crc kubenswrapper[4944]: I1124 10:24:12.156527 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 10:24:12 crc kubenswrapper[4944]: I1124 10:24:12.507578 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 10:24:13 crc kubenswrapper[4944]: I1124 10:24:13.530279 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 10:24:13 crc kubenswrapper[4944]: I1124 10:24:13.531270 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 10:24:13 crc kubenswrapper[4944]: I1124 10:24:13.578270 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 10:24:13 crc kubenswrapper[4944]: I1124 10:24:13.578351 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 10:24:14 crc kubenswrapper[4944]: I1124 10:24:14.612311 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="64282688-4ab2-436d-bc65-97c100066cd2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.83:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:24:14 crc kubenswrapper[4944]: I1124 10:24:14.612311 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="64282688-4ab2-436d-bc65-97c100066cd2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.83:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:24:14 crc kubenswrapper[4944]: I1124 10:24:14.694352 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:24:14 crc kubenswrapper[4944]: I1124 10:24:14.694430 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 10:24:15 crc kubenswrapper[4944]: I1124 10:24:15.929591 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 10:24:15 crc kubenswrapper[4944]: I1124 10:24:15.931389 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 10:24:15 crc kubenswrapper[4944]: I1124 10:24:15.933388 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 10:24:15 crc kubenswrapper[4944]: I1124 10:24:15.941167 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.029900 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.029982 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-scripts\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.030058 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6126fcec-80a3-4a35-8b17-0296d7d3f276-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.030186 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.030328 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.030354 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skpnk\" (UniqueName: \"kubernetes.io/projected/6126fcec-80a3-4a35-8b17-0296d7d3f276-kube-api-access-skpnk\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.132098 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.132202 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-scripts\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.132285 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6126fcec-80a3-4a35-8b17-0296d7d3f276-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.132333 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.132360 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6126fcec-80a3-4a35-8b17-0296d7d3f276-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.132437 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.132468 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skpnk\" (UniqueName: \"kubernetes.io/projected/6126fcec-80a3-4a35-8b17-0296d7d3f276-kube-api-access-skpnk\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.139189 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.141350 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.141748 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.145844 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-scripts\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.149629 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skpnk\" (UniqueName: \"kubernetes.io/projected/6126fcec-80a3-4a35-8b17-0296d7d3f276-kube-api-access-skpnk\") pod \"cinder-scheduler-0\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.260837 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 10:24:16 crc kubenswrapper[4944]: I1124 10:24:16.686784 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 10:24:16 crc kubenswrapper[4944]: W1124 10:24:16.691747 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6126fcec_80a3_4a35_8b17_0296d7d3f276.slice/crio-462bc2b520c14afd82a435a1b6a5e17512c2cca13b7ef996f8bc3c71b3a5d079 WatchSource:0}: Error finding container 462bc2b520c14afd82a435a1b6a5e17512c2cca13b7ef996f8bc3c71b3a5d079: Status 404 returned error can't find the container with id 462bc2b520c14afd82a435a1b6a5e17512c2cca13b7ef996f8bc3c71b3a5d079 Nov 24 10:24:17 crc kubenswrapper[4944]: I1124 10:24:17.199855 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6126fcec-80a3-4a35-8b17-0296d7d3f276","Type":"ContainerStarted","Data":"462bc2b520c14afd82a435a1b6a5e17512c2cca13b7ef996f8bc3c71b3a5d079"} Nov 24 10:24:17 crc kubenswrapper[4944]: I1124 10:24:17.358477 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 10:24:17 crc kubenswrapper[4944]: I1124 10:24:17.358905 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f72cd798-9991-4f53-84ba-6a391217a278" containerName="cinder-api-log" containerID="cri-o://2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b" gracePeriod=30 Nov 24 10:24:17 crc kubenswrapper[4944]: I1124 10:24:17.358999 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f72cd798-9991-4f53-84ba-6a391217a278" containerName="cinder-api" containerID="cri-o://b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df" gracePeriod=30 Nov 24 10:24:17 crc kubenswrapper[4944]: I1124 10:24:17.513886 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 10:24:17 crc kubenswrapper[4944]: I1124 10:24:17.993273 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 10:24:17 crc kubenswrapper[4944]: I1124 10:24:17.994828 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:17 crc kubenswrapper[4944]: I1124 10:24:17.996562 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.007794 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084123 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084681 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084724 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084749 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084774 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084801 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084833 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084855 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084891 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrm5j\" (UniqueName: \"kubernetes.io/projected/07dbc464-82a3-4762-a75b-1b4b7a236be7-kube-api-access-lrm5j\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084917 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084941 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/07dbc464-82a3-4762-a75b-1b4b7a236be7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.084965 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.085021 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-run\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.085071 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.085112 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.085139 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.186532 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrm5j\" (UniqueName: \"kubernetes.io/projected/07dbc464-82a3-4762-a75b-1b4b7a236be7-kube-api-access-lrm5j\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.186836 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.186922 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/07dbc464-82a3-4762-a75b-1b4b7a236be7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187012 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187209 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-run\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187288 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187366 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187442 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187534 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.186933 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187641 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-run\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187610 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187612 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187793 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-sys\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187831 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187883 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187895 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187916 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187942 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.188000 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.188025 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.187803 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.188208 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.188253 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.188262 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-dev\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.188353 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/07dbc464-82a3-4762-a75b-1b4b7a236be7-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.192681 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.192785 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.196917 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.197601 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/07dbc464-82a3-4762-a75b-1b4b7a236be7-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.204589 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrm5j\" (UniqueName: \"kubernetes.io/projected/07dbc464-82a3-4762-a75b-1b4b7a236be7-kube-api-access-lrm5j\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.210874 4944 generic.go:334] "Generic (PLEG): container finished" podID="f72cd798-9991-4f53-84ba-6a391217a278" containerID="2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b" exitCode=143 Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.210935 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f72cd798-9991-4f53-84ba-6a391217a278","Type":"ContainerDied","Data":"2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b"} Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.213280 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6126fcec-80a3-4a35-8b17-0296d7d3f276","Type":"ContainerStarted","Data":"e7280220640a036da261a00ad8167e97d28ed7af3a867d513c67dec04b359dc3"} Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.213315 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6126fcec-80a3-4a35-8b17-0296d7d3f276","Type":"ContainerStarted","Data":"7b3d325908fb3be09b9dd458e99b0001089c2ef16aa428b2c9210d7857257549"} Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.215339 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07dbc464-82a3-4762-a75b-1b4b7a236be7-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"07dbc464-82a3-4762-a75b-1b4b7a236be7\") " pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.243557 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.243539128 podStartE2EDuration="3.243539128s" podCreationTimestamp="2025-11-24 10:24:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:24:18.236852656 +0000 UTC m=+5518.771293118" watchObservedRunningTime="2025-11-24 10:24:18.243539128 +0000 UTC m=+5518.777979590" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.322626 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.638632 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.652672 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.655922 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.658981 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811140 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-run\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811200 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-scripts\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811242 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811299 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811324 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-dev\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811355 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-sys\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811403 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811471 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-config-data\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811494 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stzdn\" (UniqueName: \"kubernetes.io/projected/d4885864-e680-460d-9145-edb86bfabf41-kube-api-access-stzdn\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811520 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811543 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d4885864-e680-460d-9145-edb86bfabf41-ceph\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811559 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811573 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811592 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811607 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.811627 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-lib-modules\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.913610 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-run\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.913668 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-scripts\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.913707 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.913747 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.913767 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-dev\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.913797 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-sys\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.913842 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.913936 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-config-data\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.913960 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stzdn\" (UniqueName: \"kubernetes.io/projected/d4885864-e680-460d-9145-edb86bfabf41-kube-api-access-stzdn\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.913992 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914019 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d4885864-e680-460d-9145-edb86bfabf41-ceph\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914060 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914082 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914107 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914130 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914154 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-lib-modules\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914261 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-lib-modules\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914307 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-run\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914669 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914710 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914769 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-dev\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914804 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914851 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914814 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914811 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.914902 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4885864-e680-460d-9145-edb86bfabf41-sys\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.920160 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.920186 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-scripts\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.920466 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.920933 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4885864-e680-460d-9145-edb86bfabf41-config-data\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.921465 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d4885864-e680-460d-9145-edb86bfabf41-ceph\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.932658 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stzdn\" (UniqueName: \"kubernetes.io/projected/d4885864-e680-460d-9145-edb86bfabf41-kube-api-access-stzdn\") pod \"cinder-backup-0\" (UID: \"d4885864-e680-460d-9145-edb86bfabf41\") " pod="openstack/cinder-backup-0" Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.987915 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.992730 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 10:24:18 crc kubenswrapper[4944]: I1124 10:24:18.995278 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Nov 24 10:24:19 crc kubenswrapper[4944]: I1124 10:24:19.228198 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"07dbc464-82a3-4762-a75b-1b4b7a236be7","Type":"ContainerStarted","Data":"3be617b83edf8e696ef04ac554e80d54eea0d3d198387c80e422ed11116d5fa4"} Nov 24 10:24:19 crc kubenswrapper[4944]: I1124 10:24:19.517637 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.239266 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d4885864-e680-460d-9145-edb86bfabf41","Type":"ContainerStarted","Data":"a87fd6883e65c6f940c5a6bd301cb678a7fd919c121aeee697aad5f4eee77fcc"} Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.245834 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"07dbc464-82a3-4762-a75b-1b4b7a236be7","Type":"ContainerStarted","Data":"2e2d7ccfbf5d624431d472cda202f31dee81f25a37621604251de0fef4598d23"} Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.888712 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.961938 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f72cd798-9991-4f53-84ba-6a391217a278-etc-machine-id\") pod \"f72cd798-9991-4f53-84ba-6a391217a278\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.962130 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data\") pod \"f72cd798-9991-4f53-84ba-6a391217a278\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.962238 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qthvq\" (UniqueName: \"kubernetes.io/projected/f72cd798-9991-4f53-84ba-6a391217a278-kube-api-access-qthvq\") pod \"f72cd798-9991-4f53-84ba-6a391217a278\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.962317 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f72cd798-9991-4f53-84ba-6a391217a278-logs\") pod \"f72cd798-9991-4f53-84ba-6a391217a278\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.962364 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-scripts\") pod \"f72cd798-9991-4f53-84ba-6a391217a278\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.962405 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-combined-ca-bundle\") pod \"f72cd798-9991-4f53-84ba-6a391217a278\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.962425 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data-custom\") pod \"f72cd798-9991-4f53-84ba-6a391217a278\" (UID: \"f72cd798-9991-4f53-84ba-6a391217a278\") " Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.968285 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f72cd798-9991-4f53-84ba-6a391217a278-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f72cd798-9991-4f53-84ba-6a391217a278" (UID: "f72cd798-9991-4f53-84ba-6a391217a278"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.969101 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f72cd798-9991-4f53-84ba-6a391217a278-logs" (OuterVolumeSpecName: "logs") pod "f72cd798-9991-4f53-84ba-6a391217a278" (UID: "f72cd798-9991-4f53-84ba-6a391217a278"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.973509 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f72cd798-9991-4f53-84ba-6a391217a278" (UID: "f72cd798-9991-4f53-84ba-6a391217a278"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.974479 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-scripts" (OuterVolumeSpecName: "scripts") pod "f72cd798-9991-4f53-84ba-6a391217a278" (UID: "f72cd798-9991-4f53-84ba-6a391217a278"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:20 crc kubenswrapper[4944]: I1124 10:24:20.974483 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f72cd798-9991-4f53-84ba-6a391217a278-kube-api-access-qthvq" (OuterVolumeSpecName: "kube-api-access-qthvq") pod "f72cd798-9991-4f53-84ba-6a391217a278" (UID: "f72cd798-9991-4f53-84ba-6a391217a278"). InnerVolumeSpecName "kube-api-access-qthvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.021260 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f72cd798-9991-4f53-84ba-6a391217a278" (UID: "f72cd798-9991-4f53-84ba-6a391217a278"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.047545 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data" (OuterVolumeSpecName: "config-data") pod "f72cd798-9991-4f53-84ba-6a391217a278" (UID: "f72cd798-9991-4f53-84ba-6a391217a278"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.066152 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.066199 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.066221 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.066235 4944 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f72cd798-9991-4f53-84ba-6a391217a278-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.066246 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f72cd798-9991-4f53-84ba-6a391217a278-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.066258 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qthvq\" (UniqueName: \"kubernetes.io/projected/f72cd798-9991-4f53-84ba-6a391217a278-kube-api-access-qthvq\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.066270 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f72cd798-9991-4f53-84ba-6a391217a278-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.259729 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"07dbc464-82a3-4762-a75b-1b4b7a236be7","Type":"ContainerStarted","Data":"7e2f9b8d48d36fba5e7596f96703d4930473cc2a8298c8daaf94d26e3ac26974"} Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.261463 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.263140 4944 generic.go:334] "Generic (PLEG): container finished" podID="f72cd798-9991-4f53-84ba-6a391217a278" containerID="b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df" exitCode=0 Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.263210 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f72cd798-9991-4f53-84ba-6a391217a278","Type":"ContainerDied","Data":"b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df"} Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.263252 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f72cd798-9991-4f53-84ba-6a391217a278","Type":"ContainerDied","Data":"e8f9a014164659ba22f625f20fdbeb0d3509985cc8f4698fcbb474b6d537f5fc"} Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.263276 4944 scope.go:117] "RemoveContainer" containerID="b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.263271 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.267340 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d4885864-e680-460d-9145-edb86bfabf41","Type":"ContainerStarted","Data":"b159c2362e9a485475a72a168c1d541f791adb0f8521f4cdb4d14c62bbbecf04"} Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.267378 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d4885864-e680-460d-9145-edb86bfabf41","Type":"ContainerStarted","Data":"b5b6afe3649aa29820f8cc837ea4b31a5192bb8ae2b31cb5228c300f8e410e11"} Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.303348 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.507026312 podStartE2EDuration="4.303327417s" podCreationTimestamp="2025-11-24 10:24:17 +0000 UTC" firstStartedPulling="2025-11-24 10:24:18.992547448 +0000 UTC m=+5519.526987910" lastFinishedPulling="2025-11-24 10:24:19.788848553 +0000 UTC m=+5520.323289015" observedRunningTime="2025-11-24 10:24:21.28456231 +0000 UTC m=+5521.819002782" watchObservedRunningTime="2025-11-24 10:24:21.303327417 +0000 UTC m=+5521.837767879" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.320884 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.67090226 podStartE2EDuration="3.320867676s" podCreationTimestamp="2025-11-24 10:24:18 +0000 UTC" firstStartedPulling="2025-11-24 10:24:19.530131886 +0000 UTC m=+5520.064572348" lastFinishedPulling="2025-11-24 10:24:20.180097312 +0000 UTC m=+5520.714537764" observedRunningTime="2025-11-24 10:24:21.313966257 +0000 UTC m=+5521.848406739" watchObservedRunningTime="2025-11-24 10:24:21.320867676 +0000 UTC m=+5521.855308138" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.323963 4944 scope.go:117] "RemoveContainer" containerID="2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.343904 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.360477 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.399071 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 10:24:21 crc kubenswrapper[4944]: E1124 10:24:21.399835 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f72cd798-9991-4f53-84ba-6a391217a278" containerName="cinder-api" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.399847 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f72cd798-9991-4f53-84ba-6a391217a278" containerName="cinder-api" Nov 24 10:24:21 crc kubenswrapper[4944]: E1124 10:24:21.399876 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f72cd798-9991-4f53-84ba-6a391217a278" containerName="cinder-api-log" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.399882 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f72cd798-9991-4f53-84ba-6a391217a278" containerName="cinder-api-log" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.400059 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f72cd798-9991-4f53-84ba-6a391217a278" containerName="cinder-api-log" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.400074 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f72cd798-9991-4f53-84ba-6a391217a278" containerName="cinder-api" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.401085 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.405528 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.406811 4944 scope.go:117] "RemoveContainer" containerID="b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.417178 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 10:24:21 crc kubenswrapper[4944]: E1124 10:24:21.423725 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df\": container with ID starting with b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df not found: ID does not exist" containerID="b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.423760 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df"} err="failed to get container status \"b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df\": rpc error: code = NotFound desc = could not find container \"b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df\": container with ID starting with b418bb1b81a631458ec818adbf329819343da8f59b7b7990ab188bc6f62399df not found: ID does not exist" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.423784 4944 scope.go:117] "RemoveContainer" containerID="2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b" Nov 24 10:24:21 crc kubenswrapper[4944]: E1124 10:24:21.425247 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b\": container with ID starting with 2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b not found: ID does not exist" containerID="2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.425310 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b"} err="failed to get container status \"2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b\": rpc error: code = NotFound desc = could not find container \"2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b\": container with ID starting with 2185103e1b5efddeebc361ec1b3abe7dfd0217da6dca43772bc77c6c9b09d31b not found: ID does not exist" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.472444 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25q8z\" (UniqueName: \"kubernetes.io/projected/94910134-a6a9-4b2c-b29d-24098a44d4a3-kube-api-access-25q8z\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.472542 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.472582 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-scripts\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.472604 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-config-data-custom\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.472624 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94910134-a6a9-4b2c-b29d-24098a44d4a3-logs\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.472651 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-config-data\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.472685 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94910134-a6a9-4b2c-b29d-24098a44d4a3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.574353 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-scripts\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.574442 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-config-data-custom\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.574473 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94910134-a6a9-4b2c-b29d-24098a44d4a3-logs\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.574512 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-config-data\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.574564 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94910134-a6a9-4b2c-b29d-24098a44d4a3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.574630 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25q8z\" (UniqueName: \"kubernetes.io/projected/94910134-a6a9-4b2c-b29d-24098a44d4a3-kube-api-access-25q8z\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.574706 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.575012 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/94910134-a6a9-4b2c-b29d-24098a44d4a3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.576267 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94910134-a6a9-4b2c-b29d-24098a44d4a3-logs\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.579207 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-config-data-custom\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.587575 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-scripts\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.587855 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.590412 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94910134-a6a9-4b2c-b29d-24098a44d4a3-config-data\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.596160 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25q8z\" (UniqueName: \"kubernetes.io/projected/94910134-a6a9-4b2c-b29d-24098a44d4a3-kube-api-access-25q8z\") pod \"cinder-api-0\" (UID: \"94910134-a6a9-4b2c-b29d-24098a44d4a3\") " pod="openstack/cinder-api-0" Nov 24 10:24:21 crc kubenswrapper[4944]: I1124 10:24:21.749390 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 10:24:22 crc kubenswrapper[4944]: W1124 10:24:22.184384 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94910134_a6a9_4b2c_b29d_24098a44d4a3.slice/crio-a808483a23f67ce3b9005745df5135eaa3154099d94f0119fb6c47d9140d6285 WatchSource:0}: Error finding container a808483a23f67ce3b9005745df5135eaa3154099d94f0119fb6c47d9140d6285: Status 404 returned error can't find the container with id a808483a23f67ce3b9005745df5135eaa3154099d94f0119fb6c47d9140d6285 Nov 24 10:24:22 crc kubenswrapper[4944]: I1124 10:24:22.186526 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 10:24:22 crc kubenswrapper[4944]: I1124 10:24:22.292934 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f72cd798-9991-4f53-84ba-6a391217a278" path="/var/lib/kubelet/pods/f72cd798-9991-4f53-84ba-6a391217a278/volumes" Nov 24 10:24:22 crc kubenswrapper[4944]: I1124 10:24:22.293867 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"94910134-a6a9-4b2c-b29d-24098a44d4a3","Type":"ContainerStarted","Data":"a808483a23f67ce3b9005745df5135eaa3154099d94f0119fb6c47d9140d6285"} Nov 24 10:24:23 crc kubenswrapper[4944]: I1124 10:24:23.295278 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"94910134-a6a9-4b2c-b29d-24098a44d4a3","Type":"ContainerStarted","Data":"6c3504a307105fc9e4f23ee72796f23657b3f0f4a84bea02eb73edebd43afa15"} Nov 24 10:24:23 crc kubenswrapper[4944]: I1124 10:24:23.323424 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:23 crc kubenswrapper[4944]: I1124 10:24:23.537208 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 10:24:23 crc kubenswrapper[4944]: I1124 10:24:23.537651 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 10:24:23 crc kubenswrapper[4944]: I1124 10:24:23.540140 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 10:24:23 crc kubenswrapper[4944]: I1124 10:24:23.575494 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 10:24:23 crc kubenswrapper[4944]: I1124 10:24:23.582127 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 10:24:23 crc kubenswrapper[4944]: I1124 10:24:23.583476 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 10:24:23 crc kubenswrapper[4944]: I1124 10:24:23.584218 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 10:24:23 crc kubenswrapper[4944]: I1124 10:24:23.995662 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Nov 24 10:24:24 crc kubenswrapper[4944]: I1124 10:24:24.307513 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"94910134-a6a9-4b2c-b29d-24098a44d4a3","Type":"ContainerStarted","Data":"24ce5fc2aebfd9ad68ef0a7a0f7bef6d28923ed2441db9c323cae7633ad14fe2"} Nov 24 10:24:24 crc kubenswrapper[4944]: I1124 10:24:24.309313 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 10:24:24 crc kubenswrapper[4944]: I1124 10:24:24.310598 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 10:24:24 crc kubenswrapper[4944]: I1124 10:24:24.311931 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 10:24:24 crc kubenswrapper[4944]: I1124 10:24:24.335410 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.335391993 podStartE2EDuration="3.335391993s" podCreationTimestamp="2025-11-24 10:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:24:24.327455421 +0000 UTC m=+5524.861895883" watchObservedRunningTime="2025-11-24 10:24:24.335391993 +0000 UTC m=+5524.869832455" Nov 24 10:24:25 crc kubenswrapper[4944]: I1124 10:24:25.316426 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 10:24:26 crc kubenswrapper[4944]: I1124 10:24:26.469665 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 10:24:26 crc kubenswrapper[4944]: I1124 10:24:26.513531 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 10:24:27 crc kubenswrapper[4944]: I1124 10:24:27.332792 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6126fcec-80a3-4a35-8b17-0296d7d3f276" containerName="cinder-scheduler" containerID="cri-o://7b3d325908fb3be09b9dd458e99b0001089c2ef16aa428b2c9210d7857257549" gracePeriod=30 Nov 24 10:24:27 crc kubenswrapper[4944]: I1124 10:24:27.332889 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6126fcec-80a3-4a35-8b17-0296d7d3f276" containerName="probe" containerID="cri-o://e7280220640a036da261a00ad8167e97d28ed7af3a867d513c67dec04b359dc3" gracePeriod=30 Nov 24 10:24:28 crc kubenswrapper[4944]: I1124 10:24:28.347188 4944 generic.go:334] "Generic (PLEG): container finished" podID="6126fcec-80a3-4a35-8b17-0296d7d3f276" containerID="e7280220640a036da261a00ad8167e97d28ed7af3a867d513c67dec04b359dc3" exitCode=0 Nov 24 10:24:28 crc kubenswrapper[4944]: I1124 10:24:28.347336 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6126fcec-80a3-4a35-8b17-0296d7d3f276","Type":"ContainerDied","Data":"e7280220640a036da261a00ad8167e97d28ed7af3a867d513c67dec04b359dc3"} Nov 24 10:24:28 crc kubenswrapper[4944]: I1124 10:24:28.672850 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.212697 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.364516 4944 generic.go:334] "Generic (PLEG): container finished" podID="6126fcec-80a3-4a35-8b17-0296d7d3f276" containerID="7b3d325908fb3be09b9dd458e99b0001089c2ef16aa428b2c9210d7857257549" exitCode=0 Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.364561 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6126fcec-80a3-4a35-8b17-0296d7d3f276","Type":"ContainerDied","Data":"7b3d325908fb3be09b9dd458e99b0001089c2ef16aa428b2c9210d7857257549"} Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.364592 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6126fcec-80a3-4a35-8b17-0296d7d3f276","Type":"ContainerDied","Data":"462bc2b520c14afd82a435a1b6a5e17512c2cca13b7ef996f8bc3c71b3a5d079"} Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.364607 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="462bc2b520c14afd82a435a1b6a5e17512c2cca13b7ef996f8bc3c71b3a5d079" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.399662 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.556955 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-scripts\") pod \"6126fcec-80a3-4a35-8b17-0296d7d3f276\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.557037 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data\") pod \"6126fcec-80a3-4a35-8b17-0296d7d3f276\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.557106 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data-custom\") pod \"6126fcec-80a3-4a35-8b17-0296d7d3f276\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.557130 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6126fcec-80a3-4a35-8b17-0296d7d3f276-etc-machine-id\") pod \"6126fcec-80a3-4a35-8b17-0296d7d3f276\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.557234 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-combined-ca-bundle\") pod \"6126fcec-80a3-4a35-8b17-0296d7d3f276\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.557362 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skpnk\" (UniqueName: \"kubernetes.io/projected/6126fcec-80a3-4a35-8b17-0296d7d3f276-kube-api-access-skpnk\") pod \"6126fcec-80a3-4a35-8b17-0296d7d3f276\" (UID: \"6126fcec-80a3-4a35-8b17-0296d7d3f276\") " Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.558235 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6126fcec-80a3-4a35-8b17-0296d7d3f276-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6126fcec-80a3-4a35-8b17-0296d7d3f276" (UID: "6126fcec-80a3-4a35-8b17-0296d7d3f276"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.568699 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6126fcec-80a3-4a35-8b17-0296d7d3f276-kube-api-access-skpnk" (OuterVolumeSpecName: "kube-api-access-skpnk") pod "6126fcec-80a3-4a35-8b17-0296d7d3f276" (UID: "6126fcec-80a3-4a35-8b17-0296d7d3f276"). InnerVolumeSpecName "kube-api-access-skpnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.568723 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-scripts" (OuterVolumeSpecName: "scripts") pod "6126fcec-80a3-4a35-8b17-0296d7d3f276" (UID: "6126fcec-80a3-4a35-8b17-0296d7d3f276"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.568800 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6126fcec-80a3-4a35-8b17-0296d7d3f276" (UID: "6126fcec-80a3-4a35-8b17-0296d7d3f276"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.606931 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6126fcec-80a3-4a35-8b17-0296d7d3f276" (UID: "6126fcec-80a3-4a35-8b17-0296d7d3f276"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.659216 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skpnk\" (UniqueName: \"kubernetes.io/projected/6126fcec-80a3-4a35-8b17-0296d7d3f276-kube-api-access-skpnk\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.659251 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.659261 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.659270 4944 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6126fcec-80a3-4a35-8b17-0296d7d3f276-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.659279 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.664711 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data" (OuterVolumeSpecName: "config-data") pod "6126fcec-80a3-4a35-8b17-0296d7d3f276" (UID: "6126fcec-80a3-4a35-8b17-0296d7d3f276"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:24:29 crc kubenswrapper[4944]: I1124 10:24:29.761231 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6126fcec-80a3-4a35-8b17-0296d7d3f276-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.374290 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.398519 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.417377 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.425596 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 10:24:30 crc kubenswrapper[4944]: E1124 10:24:30.426107 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6126fcec-80a3-4a35-8b17-0296d7d3f276" containerName="probe" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.426132 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6126fcec-80a3-4a35-8b17-0296d7d3f276" containerName="probe" Nov 24 10:24:30 crc kubenswrapper[4944]: E1124 10:24:30.426175 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6126fcec-80a3-4a35-8b17-0296d7d3f276" containerName="cinder-scheduler" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.426183 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6126fcec-80a3-4a35-8b17-0296d7d3f276" containerName="cinder-scheduler" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.426393 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6126fcec-80a3-4a35-8b17-0296d7d3f276" containerName="cinder-scheduler" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.426425 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6126fcec-80a3-4a35-8b17-0296d7d3f276" containerName="probe" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.431300 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.432939 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.433371 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.579019 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4btb\" (UniqueName: \"kubernetes.io/projected/2e92855b-657c-4b26-a3b0-6916f4f2cb81-kube-api-access-x4btb\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.579776 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.579927 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e92855b-657c-4b26-a3b0-6916f4f2cb81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.580034 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.580232 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.580360 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.683368 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.683498 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.683555 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4btb\" (UniqueName: \"kubernetes.io/projected/2e92855b-657c-4b26-a3b0-6916f4f2cb81-kube-api-access-x4btb\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.683578 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.683649 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e92855b-657c-4b26-a3b0-6916f4f2cb81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.683690 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.684199 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e92855b-657c-4b26-a3b0-6916f4f2cb81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.691126 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.691698 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.693166 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.696336 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e92855b-657c-4b26-a3b0-6916f4f2cb81-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.700537 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4btb\" (UniqueName: \"kubernetes.io/projected/2e92855b-657c-4b26-a3b0-6916f4f2cb81-kube-api-access-x4btb\") pod \"cinder-scheduler-0\" (UID: \"2e92855b-657c-4b26-a3b0-6916f4f2cb81\") " pod="openstack/cinder-scheduler-0" Nov 24 10:24:30 crc kubenswrapper[4944]: I1124 10:24:30.753525 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 10:24:31 crc kubenswrapper[4944]: I1124 10:24:31.202768 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 10:24:31 crc kubenswrapper[4944]: W1124 10:24:31.204177 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e92855b_657c_4b26_a3b0_6916f4f2cb81.slice/crio-de5c759cd9304b9c64e1c7f6efe3c43d8f8ee62a000dda0403c927c46e0a4a04 WatchSource:0}: Error finding container de5c759cd9304b9c64e1c7f6efe3c43d8f8ee62a000dda0403c927c46e0a4a04: Status 404 returned error can't find the container with id de5c759cd9304b9c64e1c7f6efe3c43d8f8ee62a000dda0403c927c46e0a4a04 Nov 24 10:24:31 crc kubenswrapper[4944]: I1124 10:24:31.385352 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e92855b-657c-4b26-a3b0-6916f4f2cb81","Type":"ContainerStarted","Data":"de5c759cd9304b9c64e1c7f6efe3c43d8f8ee62a000dda0403c927c46e0a4a04"} Nov 24 10:24:32 crc kubenswrapper[4944]: I1124 10:24:32.287216 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6126fcec-80a3-4a35-8b17-0296d7d3f276" path="/var/lib/kubelet/pods/6126fcec-80a3-4a35-8b17-0296d7d3f276/volumes" Nov 24 10:24:32 crc kubenswrapper[4944]: I1124 10:24:32.396532 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e92855b-657c-4b26-a3b0-6916f4f2cb81","Type":"ContainerStarted","Data":"fdec1260a24e894eb606e8851b23e59e72f416a790fba96608785d56dbc5a533"} Nov 24 10:24:32 crc kubenswrapper[4944]: I1124 10:24:32.396580 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e92855b-657c-4b26-a3b0-6916f4f2cb81","Type":"ContainerStarted","Data":"bbd43dfd4d23bfba733635716552b8d39709ba48fe901d7e1c1645c095312e76"} Nov 24 10:24:33 crc kubenswrapper[4944]: I1124 10:24:33.588341 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 10:24:33 crc kubenswrapper[4944]: I1124 10:24:33.619515 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.6194819860000003 podStartE2EDuration="3.619481986s" podCreationTimestamp="2025-11-24 10:24:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:24:32.418954858 +0000 UTC m=+5532.953395340" watchObservedRunningTime="2025-11-24 10:24:33.619481986 +0000 UTC m=+5534.153922488" Nov 24 10:24:35 crc kubenswrapper[4944]: I1124 10:24:35.755099 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 10:24:40 crc kubenswrapper[4944]: I1124 10:24:40.974180 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 10:25:53 crc kubenswrapper[4944]: I1124 10:25:53.548526 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:25:53 crc kubenswrapper[4944]: I1124 10:25:53.549329 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:26:20 crc kubenswrapper[4944]: I1124 10:26:20.906616 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-dlb49"] Nov 24 10:26:20 crc kubenswrapper[4944]: I1124 10:26:20.908458 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dlb49" Nov 24 10:26:20 crc kubenswrapper[4944]: I1124 10:26:20.910395 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 10:26:20 crc kubenswrapper[4944]: I1124 10:26:20.910595 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-nj5tq" Nov 24 10:26:20 crc kubenswrapper[4944]: I1124 10:26:20.923080 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dlb49"] Nov 24 10:26:20 crc kubenswrapper[4944]: I1124 10:26:20.965826 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-ps6tm"] Nov 24 10:26:20 crc kubenswrapper[4944]: I1124 10:26:20.967819 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:20 crc kubenswrapper[4944]: I1124 10:26:20.981874 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-ps6tm"] Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.074164 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eec5069e-41ab-4c69-b297-7b2309dc8099-var-run\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.074226 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5sqf\" (UniqueName: \"kubernetes.io/projected/4fbacbdb-4237-4418-8bf4-54fefad49124-kube-api-access-x5sqf\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.074313 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eec5069e-41ab-4c69-b297-7b2309dc8099-var-log-ovn\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.074495 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-var-run\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.074644 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4fbacbdb-4237-4418-8bf4-54fefad49124-scripts\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.074747 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-etc-ovs\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.074795 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eec5069e-41ab-4c69-b297-7b2309dc8099-scripts\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.074814 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-var-lib\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.074834 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45qm5\" (UniqueName: \"kubernetes.io/projected/eec5069e-41ab-4c69-b297-7b2309dc8099-kube-api-access-45qm5\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.074912 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-var-log\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.074958 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eec5069e-41ab-4c69-b297-7b2309dc8099-var-run-ovn\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177422 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eec5069e-41ab-4c69-b297-7b2309dc8099-scripts\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177480 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-var-lib\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177506 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45qm5\" (UniqueName: \"kubernetes.io/projected/eec5069e-41ab-4c69-b297-7b2309dc8099-kube-api-access-45qm5\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177543 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-var-log\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177573 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eec5069e-41ab-4c69-b297-7b2309dc8099-var-run-ovn\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177625 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eec5069e-41ab-4c69-b297-7b2309dc8099-var-run\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177648 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5sqf\" (UniqueName: \"kubernetes.io/projected/4fbacbdb-4237-4418-8bf4-54fefad49124-kube-api-access-x5sqf\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177708 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eec5069e-41ab-4c69-b297-7b2309dc8099-var-log-ovn\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177790 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-var-run\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177847 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4fbacbdb-4237-4418-8bf4-54fefad49124-scripts\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177892 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-var-log\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177895 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eec5069e-41ab-4c69-b297-7b2309dc8099-var-run-ovn\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177896 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-etc-ovs\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177913 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-var-lib\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177906 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eec5069e-41ab-4c69-b297-7b2309dc8099-var-run\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177950 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-var-run\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177970 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4fbacbdb-4237-4418-8bf4-54fefad49124-etc-ovs\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.177968 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eec5069e-41ab-4c69-b297-7b2309dc8099-var-log-ovn\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.179908 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4fbacbdb-4237-4418-8bf4-54fefad49124-scripts\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.179925 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eec5069e-41ab-4c69-b297-7b2309dc8099-scripts\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.196623 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45qm5\" (UniqueName: \"kubernetes.io/projected/eec5069e-41ab-4c69-b297-7b2309dc8099-kube-api-access-45qm5\") pod \"ovn-controller-dlb49\" (UID: \"eec5069e-41ab-4c69-b297-7b2309dc8099\") " pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.197764 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5sqf\" (UniqueName: \"kubernetes.io/projected/4fbacbdb-4237-4418-8bf4-54fefad49124-kube-api-access-x5sqf\") pod \"ovn-controller-ovs-ps6tm\" (UID: \"4fbacbdb-4237-4418-8bf4-54fefad49124\") " pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.231738 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dlb49" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.293289 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:21 crc kubenswrapper[4944]: I1124 10:26:21.752924 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dlb49"] Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.203713 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-ps6tm"] Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.429762 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dlb49" event={"ID":"eec5069e-41ab-4c69-b297-7b2309dc8099","Type":"ContainerStarted","Data":"7b7f78554e226778de9e790ccc383d677e65009ccdcfac17110aa3b7cc02cfe1"} Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.470575 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dlb49" event={"ID":"eec5069e-41ab-4c69-b297-7b2309dc8099","Type":"ContainerStarted","Data":"14d387626201baa0afd5fe1002a6b1dbfa9c0516948480b6fb7caf25820327eb"} Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.470630 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-dlb49" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.472766 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ps6tm" event={"ID":"4fbacbdb-4237-4418-8bf4-54fefad49124","Type":"ContainerStarted","Data":"6a42f08401b86536f6c9bde8fa8aa514fd503a2707c383870c7ecc3fdca2be13"} Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.476284 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-dlb49" podStartSLOduration=2.476271239 podStartE2EDuration="2.476271239s" podCreationTimestamp="2025-11-24 10:26:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:26:22.474720989 +0000 UTC m=+5643.009161451" watchObservedRunningTime="2025-11-24 10:26:22.476271239 +0000 UTC m=+5643.010711701" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.598980 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-cx2r8"] Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.602540 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.608829 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.624760 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-cx2r8"] Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.702401 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2zjt\" (UniqueName: \"kubernetes.io/projected/33be9ed2-a523-499b-86c5-22d3953fe9dd-kube-api-access-c2zjt\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.702536 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33be9ed2-a523-499b-86c5-22d3953fe9dd-config\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.702580 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/33be9ed2-a523-499b-86c5-22d3953fe9dd-ovn-rundir\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.702619 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/33be9ed2-a523-499b-86c5-22d3953fe9dd-ovs-rundir\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.804604 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33be9ed2-a523-499b-86c5-22d3953fe9dd-config\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.805276 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/33be9ed2-a523-499b-86c5-22d3953fe9dd-ovn-rundir\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.805389 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/33be9ed2-a523-499b-86c5-22d3953fe9dd-ovs-rundir\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.805554 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2zjt\" (UniqueName: \"kubernetes.io/projected/33be9ed2-a523-499b-86c5-22d3953fe9dd-kube-api-access-c2zjt\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.805614 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33be9ed2-a523-499b-86c5-22d3953fe9dd-config\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.805768 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/33be9ed2-a523-499b-86c5-22d3953fe9dd-ovs-rundir\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.805769 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/33be9ed2-a523-499b-86c5-22d3953fe9dd-ovn-rundir\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.822970 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2zjt\" (UniqueName: \"kubernetes.io/projected/33be9ed2-a523-499b-86c5-22d3953fe9dd-kube-api-access-c2zjt\") pod \"ovn-controller-metrics-cx2r8\" (UID: \"33be9ed2-a523-499b-86c5-22d3953fe9dd\") " pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:22 crc kubenswrapper[4944]: I1124 10:26:22.955562 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-cx2r8" Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.120903 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-ztssr"] Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.122578 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ztssr" Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.134744 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-ztssr"] Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.213094 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08a84af0-675d-48d4-a3e4-d6f42781af31-operator-scripts\") pod \"octavia-db-create-ztssr\" (UID: \"08a84af0-675d-48d4-a3e4-d6f42781af31\") " pod="openstack/octavia-db-create-ztssr" Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.213218 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfdgv\" (UniqueName: \"kubernetes.io/projected/08a84af0-675d-48d4-a3e4-d6f42781af31-kube-api-access-wfdgv\") pod \"octavia-db-create-ztssr\" (UID: \"08a84af0-675d-48d4-a3e4-d6f42781af31\") " pod="openstack/octavia-db-create-ztssr" Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.314817 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08a84af0-675d-48d4-a3e4-d6f42781af31-operator-scripts\") pod \"octavia-db-create-ztssr\" (UID: \"08a84af0-675d-48d4-a3e4-d6f42781af31\") " pod="openstack/octavia-db-create-ztssr" Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.315635 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08a84af0-675d-48d4-a3e4-d6f42781af31-operator-scripts\") pod \"octavia-db-create-ztssr\" (UID: \"08a84af0-675d-48d4-a3e4-d6f42781af31\") " pod="openstack/octavia-db-create-ztssr" Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.315820 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfdgv\" (UniqueName: \"kubernetes.io/projected/08a84af0-675d-48d4-a3e4-d6f42781af31-kube-api-access-wfdgv\") pod \"octavia-db-create-ztssr\" (UID: \"08a84af0-675d-48d4-a3e4-d6f42781af31\") " pod="openstack/octavia-db-create-ztssr" Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.363379 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfdgv\" (UniqueName: \"kubernetes.io/projected/08a84af0-675d-48d4-a3e4-d6f42781af31-kube-api-access-wfdgv\") pod \"octavia-db-create-ztssr\" (UID: \"08a84af0-675d-48d4-a3e4-d6f42781af31\") " pod="openstack/octavia-db-create-ztssr" Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.440287 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ztssr" Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.549790 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.550183 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.578779 4944 generic.go:334] "Generic (PLEG): container finished" podID="4fbacbdb-4237-4418-8bf4-54fefad49124" containerID="b583586fd1022f63768eb957a76b2c9624790ae1cfafd3360f4838b14218fb6f" exitCode=0 Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.579871 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ps6tm" event={"ID":"4fbacbdb-4237-4418-8bf4-54fefad49124","Type":"ContainerDied","Data":"b583586fd1022f63768eb957a76b2c9624790ae1cfafd3360f4838b14218fb6f"} Nov 24 10:26:23 crc kubenswrapper[4944]: I1124 10:26:23.587859 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-cx2r8"] Nov 24 10:26:24 crc kubenswrapper[4944]: I1124 10:26:24.078526 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-ztssr"] Nov 24 10:26:24 crc kubenswrapper[4944]: W1124 10:26:24.091087 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08a84af0_675d_48d4_a3e4_d6f42781af31.slice/crio-b9b74d3e1f3d9b3f3a922d27164da7238997fbb77586dfa1dcbe09a3a2ea8cbb WatchSource:0}: Error finding container b9b74d3e1f3d9b3f3a922d27164da7238997fbb77586dfa1dcbe09a3a2ea8cbb: Status 404 returned error can't find the container with id b9b74d3e1f3d9b3f3a922d27164da7238997fbb77586dfa1dcbe09a3a2ea8cbb Nov 24 10:26:24 crc kubenswrapper[4944]: I1124 10:26:24.590207 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ps6tm" event={"ID":"4fbacbdb-4237-4418-8bf4-54fefad49124","Type":"ContainerStarted","Data":"435b6ea6263c58354e77bc6f4becf0bca487e265287a60200bf1b94df6d0f425"} Nov 24 10:26:24 crc kubenswrapper[4944]: I1124 10:26:24.590651 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ps6tm" event={"ID":"4fbacbdb-4237-4418-8bf4-54fefad49124","Type":"ContainerStarted","Data":"33af8ef1bc0d20261e5f5490c712b2e48d49a07d633eed8e6742d0d3bf1dd29e"} Nov 24 10:26:24 crc kubenswrapper[4944]: I1124 10:26:24.590779 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:24 crc kubenswrapper[4944]: I1124 10:26:24.592771 4944 generic.go:334] "Generic (PLEG): container finished" podID="08a84af0-675d-48d4-a3e4-d6f42781af31" containerID="0b49c7f48cc2e8d3424d50a15800d94df3f6da14722928e9e85d00d6384c5a34" exitCode=0 Nov 24 10:26:24 crc kubenswrapper[4944]: I1124 10:26:24.592834 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ztssr" event={"ID":"08a84af0-675d-48d4-a3e4-d6f42781af31","Type":"ContainerDied","Data":"0b49c7f48cc2e8d3424d50a15800d94df3f6da14722928e9e85d00d6384c5a34"} Nov 24 10:26:24 crc kubenswrapper[4944]: I1124 10:26:24.592857 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ztssr" event={"ID":"08a84af0-675d-48d4-a3e4-d6f42781af31","Type":"ContainerStarted","Data":"b9b74d3e1f3d9b3f3a922d27164da7238997fbb77586dfa1dcbe09a3a2ea8cbb"} Nov 24 10:26:24 crc kubenswrapper[4944]: I1124 10:26:24.595084 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-cx2r8" event={"ID":"33be9ed2-a523-499b-86c5-22d3953fe9dd","Type":"ContainerStarted","Data":"3defa227a2f4ecca0452fe389bd92dc400562700b497ad449a8f3643020cc7ea"} Nov 24 10:26:24 crc kubenswrapper[4944]: I1124 10:26:24.595116 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-cx2r8" event={"ID":"33be9ed2-a523-499b-86c5-22d3953fe9dd","Type":"ContainerStarted","Data":"cd1feda598ad8385f9fa5a57691f82f75cac9654458e7d51f5a4b42e685796d6"} Nov 24 10:26:24 crc kubenswrapper[4944]: I1124 10:26:24.616179 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-ps6tm" podStartSLOduration=4.616154185 podStartE2EDuration="4.616154185s" podCreationTimestamp="2025-11-24 10:26:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:26:24.611409735 +0000 UTC m=+5645.145850197" watchObservedRunningTime="2025-11-24 10:26:24.616154185 +0000 UTC m=+5645.150594647" Nov 24 10:26:24 crc kubenswrapper[4944]: I1124 10:26:24.631713 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-cx2r8" podStartSLOduration=2.631694831 podStartE2EDuration="2.631694831s" podCreationTimestamp="2025-11-24 10:26:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:26:24.62697146 +0000 UTC m=+5645.161411932" watchObservedRunningTime="2025-11-24 10:26:24.631694831 +0000 UTC m=+5645.166135293" Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.189715 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-9ac7-account-create-84fwr"] Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.191674 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-9ac7-account-create-84fwr" Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.196373 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.199261 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-9ac7-account-create-84fwr"] Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.274263 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61d0cc15-19b3-4c86-aadc-2d1814094e32-operator-scripts\") pod \"octavia-9ac7-account-create-84fwr\" (UID: \"61d0cc15-19b3-4c86-aadc-2d1814094e32\") " pod="openstack/octavia-9ac7-account-create-84fwr" Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.274444 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6gbs\" (UniqueName: \"kubernetes.io/projected/61d0cc15-19b3-4c86-aadc-2d1814094e32-kube-api-access-t6gbs\") pod \"octavia-9ac7-account-create-84fwr\" (UID: \"61d0cc15-19b3-4c86-aadc-2d1814094e32\") " pod="openstack/octavia-9ac7-account-create-84fwr" Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.376380 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6gbs\" (UniqueName: \"kubernetes.io/projected/61d0cc15-19b3-4c86-aadc-2d1814094e32-kube-api-access-t6gbs\") pod \"octavia-9ac7-account-create-84fwr\" (UID: \"61d0cc15-19b3-4c86-aadc-2d1814094e32\") " pod="openstack/octavia-9ac7-account-create-84fwr" Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.376476 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61d0cc15-19b3-4c86-aadc-2d1814094e32-operator-scripts\") pod \"octavia-9ac7-account-create-84fwr\" (UID: \"61d0cc15-19b3-4c86-aadc-2d1814094e32\") " pod="openstack/octavia-9ac7-account-create-84fwr" Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.377604 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61d0cc15-19b3-4c86-aadc-2d1814094e32-operator-scripts\") pod \"octavia-9ac7-account-create-84fwr\" (UID: \"61d0cc15-19b3-4c86-aadc-2d1814094e32\") " pod="openstack/octavia-9ac7-account-create-84fwr" Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.400963 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6gbs\" (UniqueName: \"kubernetes.io/projected/61d0cc15-19b3-4c86-aadc-2d1814094e32-kube-api-access-t6gbs\") pod \"octavia-9ac7-account-create-84fwr\" (UID: \"61d0cc15-19b3-4c86-aadc-2d1814094e32\") " pod="openstack/octavia-9ac7-account-create-84fwr" Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.517463 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-9ac7-account-create-84fwr" Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.612593 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:25 crc kubenswrapper[4944]: W1124 10:26:25.962173 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61d0cc15_19b3_4c86_aadc_2d1814094e32.slice/crio-ba00f04b10d05816860e14c754b4ec125552e8eff033ef25a26199a148e6e0f8 WatchSource:0}: Error finding container ba00f04b10d05816860e14c754b4ec125552e8eff033ef25a26199a148e6e0f8: Status 404 returned error can't find the container with id ba00f04b10d05816860e14c754b4ec125552e8eff033ef25a26199a148e6e0f8 Nov 24 10:26:25 crc kubenswrapper[4944]: I1124 10:26:25.974398 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-9ac7-account-create-84fwr"] Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.069277 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ztssr" Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.191312 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfdgv\" (UniqueName: \"kubernetes.io/projected/08a84af0-675d-48d4-a3e4-d6f42781af31-kube-api-access-wfdgv\") pod \"08a84af0-675d-48d4-a3e4-d6f42781af31\" (UID: \"08a84af0-675d-48d4-a3e4-d6f42781af31\") " Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.191404 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08a84af0-675d-48d4-a3e4-d6f42781af31-operator-scripts\") pod \"08a84af0-675d-48d4-a3e4-d6f42781af31\" (UID: \"08a84af0-675d-48d4-a3e4-d6f42781af31\") " Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.192340 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08a84af0-675d-48d4-a3e4-d6f42781af31-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "08a84af0-675d-48d4-a3e4-d6f42781af31" (UID: "08a84af0-675d-48d4-a3e4-d6f42781af31"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.197566 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08a84af0-675d-48d4-a3e4-d6f42781af31-kube-api-access-wfdgv" (OuterVolumeSpecName: "kube-api-access-wfdgv") pod "08a84af0-675d-48d4-a3e4-d6f42781af31" (UID: "08a84af0-675d-48d4-a3e4-d6f42781af31"). InnerVolumeSpecName "kube-api-access-wfdgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.294156 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08a84af0-675d-48d4-a3e4-d6f42781af31-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.294190 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfdgv\" (UniqueName: \"kubernetes.io/projected/08a84af0-675d-48d4-a3e4-d6f42781af31-kube-api-access-wfdgv\") on node \"crc\" DevicePath \"\"" Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.619916 4944 generic.go:334] "Generic (PLEG): container finished" podID="61d0cc15-19b3-4c86-aadc-2d1814094e32" containerID="96e4207108cc6c18351e5deaefe376a84218b8730a50db7bbb94489337985754" exitCode=0 Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.619975 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-9ac7-account-create-84fwr" event={"ID":"61d0cc15-19b3-4c86-aadc-2d1814094e32","Type":"ContainerDied","Data":"96e4207108cc6c18351e5deaefe376a84218b8730a50db7bbb94489337985754"} Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.620439 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-9ac7-account-create-84fwr" event={"ID":"61d0cc15-19b3-4c86-aadc-2d1814094e32","Type":"ContainerStarted","Data":"ba00f04b10d05816860e14c754b4ec125552e8eff033ef25a26199a148e6e0f8"} Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.621924 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ztssr" event={"ID":"08a84af0-675d-48d4-a3e4-d6f42781af31","Type":"ContainerDied","Data":"b9b74d3e1f3d9b3f3a922d27164da7238997fbb77586dfa1dcbe09a3a2ea8cbb"} Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.621959 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9b74d3e1f3d9b3f3a922d27164da7238997fbb77586dfa1dcbe09a3a2ea8cbb" Nov 24 10:26:26 crc kubenswrapper[4944]: I1124 10:26:26.621956 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ztssr" Nov 24 10:26:28 crc kubenswrapper[4944]: I1124 10:26:28.043798 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-9ac7-account-create-84fwr" Nov 24 10:26:28 crc kubenswrapper[4944]: I1124 10:26:28.126419 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61d0cc15-19b3-4c86-aadc-2d1814094e32-operator-scripts\") pod \"61d0cc15-19b3-4c86-aadc-2d1814094e32\" (UID: \"61d0cc15-19b3-4c86-aadc-2d1814094e32\") " Nov 24 10:26:28 crc kubenswrapper[4944]: I1124 10:26:28.127261 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d0cc15-19b3-4c86-aadc-2d1814094e32-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "61d0cc15-19b3-4c86-aadc-2d1814094e32" (UID: "61d0cc15-19b3-4c86-aadc-2d1814094e32"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:26:28 crc kubenswrapper[4944]: I1124 10:26:28.128571 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6gbs\" (UniqueName: \"kubernetes.io/projected/61d0cc15-19b3-4c86-aadc-2d1814094e32-kube-api-access-t6gbs\") pod \"61d0cc15-19b3-4c86-aadc-2d1814094e32\" (UID: \"61d0cc15-19b3-4c86-aadc-2d1814094e32\") " Nov 24 10:26:28 crc kubenswrapper[4944]: I1124 10:26:28.129931 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/61d0cc15-19b3-4c86-aadc-2d1814094e32-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:26:28 crc kubenswrapper[4944]: I1124 10:26:28.132963 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d0cc15-19b3-4c86-aadc-2d1814094e32-kube-api-access-t6gbs" (OuterVolumeSpecName: "kube-api-access-t6gbs") pod "61d0cc15-19b3-4c86-aadc-2d1814094e32" (UID: "61d0cc15-19b3-4c86-aadc-2d1814094e32"). InnerVolumeSpecName "kube-api-access-t6gbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:26:28 crc kubenswrapper[4944]: I1124 10:26:28.231317 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6gbs\" (UniqueName: \"kubernetes.io/projected/61d0cc15-19b3-4c86-aadc-2d1814094e32-kube-api-access-t6gbs\") on node \"crc\" DevicePath \"\"" Nov 24 10:26:28 crc kubenswrapper[4944]: I1124 10:26:28.640965 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-9ac7-account-create-84fwr" event={"ID":"61d0cc15-19b3-4c86-aadc-2d1814094e32","Type":"ContainerDied","Data":"ba00f04b10d05816860e14c754b4ec125552e8eff033ef25a26199a148e6e0f8"} Nov 24 10:26:28 crc kubenswrapper[4944]: I1124 10:26:28.641009 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba00f04b10d05816860e14c754b4ec125552e8eff033ef25a26199a148e6e0f8" Nov 24 10:26:28 crc kubenswrapper[4944]: I1124 10:26:28.641025 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-9ac7-account-create-84fwr" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.150776 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-4hk8b"] Nov 24 10:26:31 crc kubenswrapper[4944]: E1124 10:26:31.151602 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d0cc15-19b3-4c86-aadc-2d1814094e32" containerName="mariadb-account-create" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.151615 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d0cc15-19b3-4c86-aadc-2d1814094e32" containerName="mariadb-account-create" Nov 24 10:26:31 crc kubenswrapper[4944]: E1124 10:26:31.151637 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08a84af0-675d-48d4-a3e4-d6f42781af31" containerName="mariadb-database-create" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.151643 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="08a84af0-675d-48d4-a3e4-d6f42781af31" containerName="mariadb-database-create" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.151824 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="08a84af0-675d-48d4-a3e4-d6f42781af31" containerName="mariadb-database-create" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.151837 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d0cc15-19b3-4c86-aadc-2d1814094e32" containerName="mariadb-account-create" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.152498 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-4hk8b" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.162911 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-4hk8b"] Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.290215 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnk27\" (UniqueName: \"kubernetes.io/projected/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-kube-api-access-lnk27\") pod \"octavia-persistence-db-create-4hk8b\" (UID: \"b7caf33c-af3c-4e3e-aba3-a7480fa69b85\") " pod="openstack/octavia-persistence-db-create-4hk8b" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.290264 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-operator-scripts\") pod \"octavia-persistence-db-create-4hk8b\" (UID: \"b7caf33c-af3c-4e3e-aba3-a7480fa69b85\") " pod="openstack/octavia-persistence-db-create-4hk8b" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.392383 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnk27\" (UniqueName: \"kubernetes.io/projected/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-kube-api-access-lnk27\") pod \"octavia-persistence-db-create-4hk8b\" (UID: \"b7caf33c-af3c-4e3e-aba3-a7480fa69b85\") " pod="openstack/octavia-persistence-db-create-4hk8b" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.392427 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-operator-scripts\") pod \"octavia-persistence-db-create-4hk8b\" (UID: \"b7caf33c-af3c-4e3e-aba3-a7480fa69b85\") " pod="openstack/octavia-persistence-db-create-4hk8b" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.393598 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-operator-scripts\") pod \"octavia-persistence-db-create-4hk8b\" (UID: \"b7caf33c-af3c-4e3e-aba3-a7480fa69b85\") " pod="openstack/octavia-persistence-db-create-4hk8b" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.415040 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnk27\" (UniqueName: \"kubernetes.io/projected/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-kube-api-access-lnk27\") pod \"octavia-persistence-db-create-4hk8b\" (UID: \"b7caf33c-af3c-4e3e-aba3-a7480fa69b85\") " pod="openstack/octavia-persistence-db-create-4hk8b" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.475156 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-4hk8b" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.553010 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-ed3e-account-create-cb28f"] Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.564651 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-ed3e-account-create-cb28f" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.567399 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.598075 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-ed3e-account-create-cb28f"] Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.698494 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6j4d\" (UniqueName: \"kubernetes.io/projected/2de39c88-cdcc-41db-9bdc-13b0d990a451-kube-api-access-t6j4d\") pod \"octavia-ed3e-account-create-cb28f\" (UID: \"2de39c88-cdcc-41db-9bdc-13b0d990a451\") " pod="openstack/octavia-ed3e-account-create-cb28f" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.699064 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2de39c88-cdcc-41db-9bdc-13b0d990a451-operator-scripts\") pod \"octavia-ed3e-account-create-cb28f\" (UID: \"2de39c88-cdcc-41db-9bdc-13b0d990a451\") " pod="openstack/octavia-ed3e-account-create-cb28f" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.801204 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2de39c88-cdcc-41db-9bdc-13b0d990a451-operator-scripts\") pod \"octavia-ed3e-account-create-cb28f\" (UID: \"2de39c88-cdcc-41db-9bdc-13b0d990a451\") " pod="openstack/octavia-ed3e-account-create-cb28f" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.801328 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6j4d\" (UniqueName: \"kubernetes.io/projected/2de39c88-cdcc-41db-9bdc-13b0d990a451-kube-api-access-t6j4d\") pod \"octavia-ed3e-account-create-cb28f\" (UID: \"2de39c88-cdcc-41db-9bdc-13b0d990a451\") " pod="openstack/octavia-ed3e-account-create-cb28f" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.802693 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2de39c88-cdcc-41db-9bdc-13b0d990a451-operator-scripts\") pod \"octavia-ed3e-account-create-cb28f\" (UID: \"2de39c88-cdcc-41db-9bdc-13b0d990a451\") " pod="openstack/octavia-ed3e-account-create-cb28f" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.821067 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6j4d\" (UniqueName: \"kubernetes.io/projected/2de39c88-cdcc-41db-9bdc-13b0d990a451-kube-api-access-t6j4d\") pod \"octavia-ed3e-account-create-cb28f\" (UID: \"2de39c88-cdcc-41db-9bdc-13b0d990a451\") " pod="openstack/octavia-ed3e-account-create-cb28f" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.924820 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-ed3e-account-create-cb28f" Nov 24 10:26:31 crc kubenswrapper[4944]: I1124 10:26:31.946274 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-4hk8b"] Nov 24 10:26:32 crc kubenswrapper[4944]: I1124 10:26:32.358303 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-ed3e-account-create-cb28f"] Nov 24 10:26:32 crc kubenswrapper[4944]: W1124 10:26:32.362222 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2de39c88_cdcc_41db_9bdc_13b0d990a451.slice/crio-7b6456991dbe4d4f866e97f311fd7c770c8264236766c80ec9d66b6e9974397e WatchSource:0}: Error finding container 7b6456991dbe4d4f866e97f311fd7c770c8264236766c80ec9d66b6e9974397e: Status 404 returned error can't find the container with id 7b6456991dbe4d4f866e97f311fd7c770c8264236766c80ec9d66b6e9974397e Nov 24 10:26:32 crc kubenswrapper[4944]: I1124 10:26:32.677939 4944 generic.go:334] "Generic (PLEG): container finished" podID="b7caf33c-af3c-4e3e-aba3-a7480fa69b85" containerID="110d451e530b0df396a8093a35d92ec7d64fd08c04ca4d482aaa619f0f3bf2ad" exitCode=0 Nov 24 10:26:32 crc kubenswrapper[4944]: I1124 10:26:32.678036 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-4hk8b" event={"ID":"b7caf33c-af3c-4e3e-aba3-a7480fa69b85","Type":"ContainerDied","Data":"110d451e530b0df396a8093a35d92ec7d64fd08c04ca4d482aaa619f0f3bf2ad"} Nov 24 10:26:32 crc kubenswrapper[4944]: I1124 10:26:32.678431 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-4hk8b" event={"ID":"b7caf33c-af3c-4e3e-aba3-a7480fa69b85","Type":"ContainerStarted","Data":"ceb208f3128e9d038a72ce42b7931284ffbab300c2c87a5649020243d31bc8d5"} Nov 24 10:26:32 crc kubenswrapper[4944]: I1124 10:26:32.680348 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-ed3e-account-create-cb28f" event={"ID":"2de39c88-cdcc-41db-9bdc-13b0d990a451","Type":"ContainerStarted","Data":"51daae17ad1be180b249965e9c850e8d7dbf2252bbbb4f1dbded9e2e35872525"} Nov 24 10:26:32 crc kubenswrapper[4944]: I1124 10:26:32.680388 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-ed3e-account-create-cb28f" event={"ID":"2de39c88-cdcc-41db-9bdc-13b0d990a451","Type":"ContainerStarted","Data":"7b6456991dbe4d4f866e97f311fd7c770c8264236766c80ec9d66b6e9974397e"} Nov 24 10:26:32 crc kubenswrapper[4944]: I1124 10:26:32.710267 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-ed3e-account-create-cb28f" podStartSLOduration=1.710249186 podStartE2EDuration="1.710249186s" podCreationTimestamp="2025-11-24 10:26:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:26:32.703412169 +0000 UTC m=+5653.237852651" watchObservedRunningTime="2025-11-24 10:26:32.710249186 +0000 UTC m=+5653.244689648" Nov 24 10:26:32 crc kubenswrapper[4944]: E1124 10:26:32.938059 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2de39c88_cdcc_41db_9bdc_13b0d990a451.slice/crio-51daae17ad1be180b249965e9c850e8d7dbf2252bbbb4f1dbded9e2e35872525.scope\": RecentStats: unable to find data in memory cache]" Nov 24 10:26:33 crc kubenswrapper[4944]: I1124 10:26:33.689713 4944 generic.go:334] "Generic (PLEG): container finished" podID="2de39c88-cdcc-41db-9bdc-13b0d990a451" containerID="51daae17ad1be180b249965e9c850e8d7dbf2252bbbb4f1dbded9e2e35872525" exitCode=0 Nov 24 10:26:33 crc kubenswrapper[4944]: I1124 10:26:33.689761 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-ed3e-account-create-cb28f" event={"ID":"2de39c88-cdcc-41db-9bdc-13b0d990a451","Type":"ContainerDied","Data":"51daae17ad1be180b249965e9c850e8d7dbf2252bbbb4f1dbded9e2e35872525"} Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.038290 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-4hk8b" Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.058278 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-nbr9v"] Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.067331 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-812f-account-create-ksj7z"] Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.114837 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-nbr9v"] Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.123841 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-812f-account-create-ksj7z"] Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.141621 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnk27\" (UniqueName: \"kubernetes.io/projected/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-kube-api-access-lnk27\") pod \"b7caf33c-af3c-4e3e-aba3-a7480fa69b85\" (UID: \"b7caf33c-af3c-4e3e-aba3-a7480fa69b85\") " Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.141836 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-operator-scripts\") pod \"b7caf33c-af3c-4e3e-aba3-a7480fa69b85\" (UID: \"b7caf33c-af3c-4e3e-aba3-a7480fa69b85\") " Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.142905 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b7caf33c-af3c-4e3e-aba3-a7480fa69b85" (UID: "b7caf33c-af3c-4e3e-aba3-a7480fa69b85"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.150284 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-kube-api-access-lnk27" (OuterVolumeSpecName: "kube-api-access-lnk27") pod "b7caf33c-af3c-4e3e-aba3-a7480fa69b85" (UID: "b7caf33c-af3c-4e3e-aba3-a7480fa69b85"). InnerVolumeSpecName "kube-api-access-lnk27". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.243553 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnk27\" (UniqueName: \"kubernetes.io/projected/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-kube-api-access-lnk27\") on node \"crc\" DevicePath \"\"" Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.243626 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7caf33c-af3c-4e3e-aba3-a7480fa69b85-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.287850 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b7743bd-f104-434a-9a4d-59db71cb9a97" path="/var/lib/kubelet/pods/4b7743bd-f104-434a-9a4d-59db71cb9a97/volumes" Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.288475 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba4b8d12-578d-4a2e-836a-09c2b86b2a9b" path="/var/lib/kubelet/pods/ba4b8d12-578d-4a2e-836a-09c2b86b2a9b/volumes" Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.700401 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-4hk8b" Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.700452 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-4hk8b" event={"ID":"b7caf33c-af3c-4e3e-aba3-a7480fa69b85","Type":"ContainerDied","Data":"ceb208f3128e9d038a72ce42b7931284ffbab300c2c87a5649020243d31bc8d5"} Nov 24 10:26:34 crc kubenswrapper[4944]: I1124 10:26:34.700485 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ceb208f3128e9d038a72ce42b7931284ffbab300c2c87a5649020243d31bc8d5" Nov 24 10:26:35 crc kubenswrapper[4944]: I1124 10:26:35.037723 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-ed3e-account-create-cb28f" Nov 24 10:26:35 crc kubenswrapper[4944]: I1124 10:26:35.158262 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6j4d\" (UniqueName: \"kubernetes.io/projected/2de39c88-cdcc-41db-9bdc-13b0d990a451-kube-api-access-t6j4d\") pod \"2de39c88-cdcc-41db-9bdc-13b0d990a451\" (UID: \"2de39c88-cdcc-41db-9bdc-13b0d990a451\") " Nov 24 10:26:35 crc kubenswrapper[4944]: I1124 10:26:35.158369 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2de39c88-cdcc-41db-9bdc-13b0d990a451-operator-scripts\") pod \"2de39c88-cdcc-41db-9bdc-13b0d990a451\" (UID: \"2de39c88-cdcc-41db-9bdc-13b0d990a451\") " Nov 24 10:26:35 crc kubenswrapper[4944]: I1124 10:26:35.159407 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2de39c88-cdcc-41db-9bdc-13b0d990a451-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2de39c88-cdcc-41db-9bdc-13b0d990a451" (UID: "2de39c88-cdcc-41db-9bdc-13b0d990a451"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:26:35 crc kubenswrapper[4944]: I1124 10:26:35.164890 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2de39c88-cdcc-41db-9bdc-13b0d990a451-kube-api-access-t6j4d" (OuterVolumeSpecName: "kube-api-access-t6j4d") pod "2de39c88-cdcc-41db-9bdc-13b0d990a451" (UID: "2de39c88-cdcc-41db-9bdc-13b0d990a451"). InnerVolumeSpecName "kube-api-access-t6j4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:26:35 crc kubenswrapper[4944]: I1124 10:26:35.260160 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6j4d\" (UniqueName: \"kubernetes.io/projected/2de39c88-cdcc-41db-9bdc-13b0d990a451-kube-api-access-t6j4d\") on node \"crc\" DevicePath \"\"" Nov 24 10:26:35 crc kubenswrapper[4944]: I1124 10:26:35.260193 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2de39c88-cdcc-41db-9bdc-13b0d990a451-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:26:35 crc kubenswrapper[4944]: I1124 10:26:35.710116 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-ed3e-account-create-cb28f" event={"ID":"2de39c88-cdcc-41db-9bdc-13b0d990a451","Type":"ContainerDied","Data":"7b6456991dbe4d4f866e97f311fd7c770c8264236766c80ec9d66b6e9974397e"} Nov 24 10:26:35 crc kubenswrapper[4944]: I1124 10:26:35.710390 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b6456991dbe4d4f866e97f311fd7c770c8264236766c80ec9d66b6e9974397e" Nov 24 10:26:35 crc kubenswrapper[4944]: I1124 10:26:35.710193 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-ed3e-account-create-cb28f" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.201965 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-75588859b8-79hh7"] Nov 24 10:26:37 crc kubenswrapper[4944]: E1124 10:26:37.202514 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de39c88-cdcc-41db-9bdc-13b0d990a451" containerName="mariadb-account-create" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.202531 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de39c88-cdcc-41db-9bdc-13b0d990a451" containerName="mariadb-account-create" Nov 24 10:26:37 crc kubenswrapper[4944]: E1124 10:26:37.202548 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7caf33c-af3c-4e3e-aba3-a7480fa69b85" containerName="mariadb-database-create" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.202557 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7caf33c-af3c-4e3e-aba3-a7480fa69b85" containerName="mariadb-database-create" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.202783 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2de39c88-cdcc-41db-9bdc-13b0d990a451" containerName="mariadb-account-create" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.202808 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7caf33c-af3c-4e3e-aba3-a7480fa69b85" containerName="mariadb-database-create" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.204489 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.207605 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.208271 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.214416 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-s9l8c" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.226508 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-75588859b8-79hh7"] Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.297450 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3012a4dd-0211-4ada-a25c-234d280caa1a-config-data\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.297749 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3012a4dd-0211-4ada-a25c-234d280caa1a-combined-ca-bundle\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.297900 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3012a4dd-0211-4ada-a25c-234d280caa1a-config-data-merged\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.298119 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3012a4dd-0211-4ada-a25c-234d280caa1a-scripts\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.300263 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/3012a4dd-0211-4ada-a25c-234d280caa1a-octavia-run\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.402507 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3012a4dd-0211-4ada-a25c-234d280caa1a-scripts\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.402561 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/3012a4dd-0211-4ada-a25c-234d280caa1a-octavia-run\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.402638 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3012a4dd-0211-4ada-a25c-234d280caa1a-config-data\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.402679 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3012a4dd-0211-4ada-a25c-234d280caa1a-combined-ca-bundle\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.402714 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3012a4dd-0211-4ada-a25c-234d280caa1a-config-data-merged\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.403315 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3012a4dd-0211-4ada-a25c-234d280caa1a-config-data-merged\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.405427 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/3012a4dd-0211-4ada-a25c-234d280caa1a-octavia-run\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.409340 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3012a4dd-0211-4ada-a25c-234d280caa1a-scripts\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.409678 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3012a4dd-0211-4ada-a25c-234d280caa1a-config-data\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.412242 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3012a4dd-0211-4ada-a25c-234d280caa1a-combined-ca-bundle\") pod \"octavia-api-75588859b8-79hh7\" (UID: \"3012a4dd-0211-4ada-a25c-234d280caa1a\") " pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:37 crc kubenswrapper[4944]: I1124 10:26:37.522278 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:38 crc kubenswrapper[4944]: W1124 10:26:38.002923 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3012a4dd_0211_4ada_a25c_234d280caa1a.slice/crio-395c3ae34d35844a8ac1cd5e3e4967c0d13e5d634606e2685607a0658ea53b92 WatchSource:0}: Error finding container 395c3ae34d35844a8ac1cd5e3e4967c0d13e5d634606e2685607a0658ea53b92: Status 404 returned error can't find the container with id 395c3ae34d35844a8ac1cd5e3e4967c0d13e5d634606e2685607a0658ea53b92 Nov 24 10:26:38 crc kubenswrapper[4944]: I1124 10:26:38.036679 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-75588859b8-79hh7"] Nov 24 10:26:38 crc kubenswrapper[4944]: I1124 10:26:38.738637 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-75588859b8-79hh7" event={"ID":"3012a4dd-0211-4ada-a25c-234d280caa1a","Type":"ContainerStarted","Data":"395c3ae34d35844a8ac1cd5e3e4967c0d13e5d634606e2685607a0658ea53b92"} Nov 24 10:26:40 crc kubenswrapper[4944]: I1124 10:26:40.051621 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-jm4kh"] Nov 24 10:26:40 crc kubenswrapper[4944]: I1124 10:26:40.062214 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-jm4kh"] Nov 24 10:26:40 crc kubenswrapper[4944]: I1124 10:26:40.292629 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9675c22-64db-452a-8d38-dd588bd5b43a" path="/var/lib/kubelet/pods/a9675c22-64db-452a-8d38-dd588bd5b43a/volumes" Nov 24 10:26:46 crc kubenswrapper[4944]: I1124 10:26:46.820006 4944 generic.go:334] "Generic (PLEG): container finished" podID="3012a4dd-0211-4ada-a25c-234d280caa1a" containerID="cb1edcd612b413df12db2da5cf5fef88e323567d3e082f1adf536d9c1e82c287" exitCode=0 Nov 24 10:26:46 crc kubenswrapper[4944]: I1124 10:26:46.821714 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-75588859b8-79hh7" event={"ID":"3012a4dd-0211-4ada-a25c-234d280caa1a","Type":"ContainerDied","Data":"cb1edcd612b413df12db2da5cf5fef88e323567d3e082f1adf536d9c1e82c287"} Nov 24 10:26:47 crc kubenswrapper[4944]: I1124 10:26:47.833114 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-75588859b8-79hh7" event={"ID":"3012a4dd-0211-4ada-a25c-234d280caa1a","Type":"ContainerStarted","Data":"7eb2e6404c3f87d26c17a3ff79e405f1174ef637f920a46182b7e9de77ac95d1"} Nov 24 10:26:47 crc kubenswrapper[4944]: I1124 10:26:47.833534 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-75588859b8-79hh7" event={"ID":"3012a4dd-0211-4ada-a25c-234d280caa1a","Type":"ContainerStarted","Data":"ce608dfb9627138cc357a86b00c3a3d94fc300231fb22cb44869f8bd6efc1b8d"} Nov 24 10:26:47 crc kubenswrapper[4944]: I1124 10:26:47.833555 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:47 crc kubenswrapper[4944]: I1124 10:26:47.833568 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:47 crc kubenswrapper[4944]: I1124 10:26:47.863352 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-75588859b8-79hh7" podStartSLOduration=2.7696366709999998 podStartE2EDuration="10.863333038s" podCreationTimestamp="2025-11-24 10:26:37 +0000 UTC" firstStartedPulling="2025-11-24 10:26:38.006055995 +0000 UTC m=+5658.540496457" lastFinishedPulling="2025-11-24 10:26:46.099752362 +0000 UTC m=+5666.634192824" observedRunningTime="2025-11-24 10:26:47.857369808 +0000 UTC m=+5668.391810290" watchObservedRunningTime="2025-11-24 10:26:47.863333038 +0000 UTC m=+5668.397773520" Nov 24 10:26:53 crc kubenswrapper[4944]: I1124 10:26:53.033219 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-pw5fm"] Nov 24 10:26:53 crc kubenswrapper[4944]: I1124 10:26:53.044611 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-pw5fm"] Nov 24 10:26:53 crc kubenswrapper[4944]: I1124 10:26:53.549579 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:26:53 crc kubenswrapper[4944]: I1124 10:26:53.549637 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:26:53 crc kubenswrapper[4944]: I1124 10:26:53.549679 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 10:26:53 crc kubenswrapper[4944]: I1124 10:26:53.550465 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 10:26:53 crc kubenswrapper[4944]: I1124 10:26:53.550518 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" gracePeriod=600 Nov 24 10:26:53 crc kubenswrapper[4944]: E1124 10:26:53.689796 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:26:53 crc kubenswrapper[4944]: I1124 10:26:53.885650 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" exitCode=0 Nov 24 10:26:53 crc kubenswrapper[4944]: I1124 10:26:53.885706 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754"} Nov 24 10:26:53 crc kubenswrapper[4944]: I1124 10:26:53.885755 4944 scope.go:117] "RemoveContainer" containerID="3b50540cdc5646c23244da47ca4ab4ee1e53c396cc0d5f9070264c91666bc8f2" Nov 24 10:26:53 crc kubenswrapper[4944]: I1124 10:26:53.886738 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:26:53 crc kubenswrapper[4944]: E1124 10:26:53.887081 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.289720 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f550d834-f1ef-41df-92c0-dec056f41d0e" path="/var/lib/kubelet/pods/f550d834-f1ef-41df-92c0-dec056f41d0e/volumes" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.602951 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-m8c9f"] Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.604502 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.606439 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.606490 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.606753 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.619832 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-m8c9f"] Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.696844 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-config-data\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.696893 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-hm-ports\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.696930 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-config-data-merged\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.696984 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-scripts\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.802750 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-scripts\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.803754 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-config-data\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.803772 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-hm-ports\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.803813 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-config-data-merged\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.804306 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-config-data-merged\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.804989 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-hm-ports\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.810379 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-config-data\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.812155 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c49dd9-e55b-44f7-9dd6-50ec44c9fde2-scripts\") pod \"octavia-rsyslog-m8c9f\" (UID: \"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2\") " pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:54 crc kubenswrapper[4944]: I1124 10:26:54.962971 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.552216 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-g572j"] Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.554836 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-g572j" Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.562965 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.581319 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-g572j"] Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.634687 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/db61872a-0b2a-4fbb-b3ee-36570be61ff6-httpd-config\") pod \"octavia-image-upload-59f8cff499-g572j\" (UID: \"db61872a-0b2a-4fbb-b3ee-36570be61ff6\") " pod="openstack/octavia-image-upload-59f8cff499-g572j" Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.634801 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/db61872a-0b2a-4fbb-b3ee-36570be61ff6-amphora-image\") pod \"octavia-image-upload-59f8cff499-g572j\" (UID: \"db61872a-0b2a-4fbb-b3ee-36570be61ff6\") " pod="openstack/octavia-image-upload-59f8cff499-g572j" Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.737429 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/db61872a-0b2a-4fbb-b3ee-36570be61ff6-httpd-config\") pod \"octavia-image-upload-59f8cff499-g572j\" (UID: \"db61872a-0b2a-4fbb-b3ee-36570be61ff6\") " pod="openstack/octavia-image-upload-59f8cff499-g572j" Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.737573 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/db61872a-0b2a-4fbb-b3ee-36570be61ff6-amphora-image\") pod \"octavia-image-upload-59f8cff499-g572j\" (UID: \"db61872a-0b2a-4fbb-b3ee-36570be61ff6\") " pod="openstack/octavia-image-upload-59f8cff499-g572j" Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.737981 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/db61872a-0b2a-4fbb-b3ee-36570be61ff6-amphora-image\") pod \"octavia-image-upload-59f8cff499-g572j\" (UID: \"db61872a-0b2a-4fbb-b3ee-36570be61ff6\") " pod="openstack/octavia-image-upload-59f8cff499-g572j" Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.743376 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/db61872a-0b2a-4fbb-b3ee-36570be61ff6-httpd-config\") pod \"octavia-image-upload-59f8cff499-g572j\" (UID: \"db61872a-0b2a-4fbb-b3ee-36570be61ff6\") " pod="openstack/octavia-image-upload-59f8cff499-g572j" Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.797114 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-m8c9f"] Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.910504 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-m8c9f" event={"ID":"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2","Type":"ContainerStarted","Data":"80ce70459d0e78a9ba88340d76a3264bbec4cbaf51208c842a3ecebf7b41fc3d"} Nov 24 10:26:55 crc kubenswrapper[4944]: I1124 10:26:55.935129 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-g572j" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.279593 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-dlb49" podUID="eec5069e-41ab-4c69-b297-7b2309dc8099" containerName="ovn-controller" probeResult="failure" output=< Nov 24 10:26:56 crc kubenswrapper[4944]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 10:26:56 crc kubenswrapper[4944]: > Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.342412 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.347612 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-ps6tm" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.447063 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-g572j"] Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.512633 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-dlb49-config-zx6fj"] Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.514036 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.518298 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.520877 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dlb49-config-zx6fj"] Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.654028 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-scripts\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.654130 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-log-ovn\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.654164 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.654189 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp6gq\" (UniqueName: \"kubernetes.io/projected/8e900a41-c222-4bb7-b8da-5d038165be5b-kube-api-access-zp6gq\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.654264 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run-ovn\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.654314 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-additional-scripts\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.756337 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.756637 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp6gq\" (UniqueName: \"kubernetes.io/projected/8e900a41-c222-4bb7-b8da-5d038165be5b-kube-api-access-zp6gq\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.756740 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run-ovn\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.756738 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.756839 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run-ovn\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.756934 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-additional-scripts\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.757030 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-scripts\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.757141 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-log-ovn\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.757403 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-log-ovn\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.758103 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-additional-scripts\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.759759 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-scripts\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.782624 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp6gq\" (UniqueName: \"kubernetes.io/projected/8e900a41-c222-4bb7-b8da-5d038165be5b-kube-api-access-zp6gq\") pod \"ovn-controller-dlb49-config-zx6fj\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.848068 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:26:56 crc kubenswrapper[4944]: I1124 10:26:56.929399 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-g572j" event={"ID":"db61872a-0b2a-4fbb-b3ee-36570be61ff6","Type":"ContainerStarted","Data":"322ac47f69dbda1a456c16051db0524aed6519f33e9b7f036c3a0fbe59e297e1"} Nov 24 10:26:57 crc kubenswrapper[4944]: I1124 10:26:57.367167 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dlb49-config-zx6fj"] Nov 24 10:26:57 crc kubenswrapper[4944]: I1124 10:26:57.469090 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:57 crc kubenswrapper[4944]: I1124 10:26:57.781880 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-75588859b8-79hh7" Nov 24 10:26:58 crc kubenswrapper[4944]: I1124 10:26:58.958175 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dlb49-config-zx6fj" event={"ID":"8e900a41-c222-4bb7-b8da-5d038165be5b","Type":"ContainerStarted","Data":"354bd92a08979d553389fc5021ea88727a5a754ce8d338bfd2d17892e3a87929"} Nov 24 10:26:59 crc kubenswrapper[4944]: I1124 10:26:59.990690 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dlb49-config-zx6fj" event={"ID":"8e900a41-c222-4bb7-b8da-5d038165be5b","Type":"ContainerStarted","Data":"f598824cf5db9267ba2299c12065aca660b33723888f50211a17bcf26b4903a6"} Nov 24 10:27:00 crc kubenswrapper[4944]: I1124 10:27:00.032541 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-dlb49-config-zx6fj" podStartSLOduration=4.032523246 podStartE2EDuration="4.032523246s" podCreationTimestamp="2025-11-24 10:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:27:00.030359426 +0000 UTC m=+5680.564799898" watchObservedRunningTime="2025-11-24 10:27:00.032523246 +0000 UTC m=+5680.566963708" Nov 24 10:27:01 crc kubenswrapper[4944]: I1124 10:27:01.002739 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-m8c9f" event={"ID":"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2","Type":"ContainerStarted","Data":"66b620876428df4270a367b339159d9fd61ec8ccffff77fa63fb68ee82330828"} Nov 24 10:27:01 crc kubenswrapper[4944]: I1124 10:27:01.005180 4944 generic.go:334] "Generic (PLEG): container finished" podID="8e900a41-c222-4bb7-b8da-5d038165be5b" containerID="f598824cf5db9267ba2299c12065aca660b33723888f50211a17bcf26b4903a6" exitCode=0 Nov 24 10:27:01 crc kubenswrapper[4944]: I1124 10:27:01.005214 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dlb49-config-zx6fj" event={"ID":"8e900a41-c222-4bb7-b8da-5d038165be5b","Type":"ContainerDied","Data":"f598824cf5db9267ba2299c12065aca660b33723888f50211a17bcf26b4903a6"} Nov 24 10:27:01 crc kubenswrapper[4944]: I1124 10:27:01.331927 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-dlb49" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.025585 4944 generic.go:334] "Generic (PLEG): container finished" podID="36c49dd9-e55b-44f7-9dd6-50ec44c9fde2" containerID="66b620876428df4270a367b339159d9fd61ec8ccffff77fa63fb68ee82330828" exitCode=0 Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.025741 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-m8c9f" event={"ID":"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2","Type":"ContainerDied","Data":"66b620876428df4270a367b339159d9fd61ec8ccffff77fa63fb68ee82330828"} Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.076618 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-hsl6x"] Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.078350 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.080174 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.089485 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-hsl6x"] Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.221072 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.221119 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-combined-ca-bundle\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.221210 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data-merged\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.221280 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-scripts\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.323290 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data-merged\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.323379 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-scripts\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.323465 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.323490 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-combined-ca-bundle\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.324919 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data-merged\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.332190 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-combined-ca-bundle\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.332620 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.335641 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-scripts\") pod \"octavia-db-sync-hsl6x\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:03 crc kubenswrapper[4944]: I1124 10:27:03.411563 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.436889 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.598229 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run-ovn\") pod \"8e900a41-c222-4bb7-b8da-5d038165be5b\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.598317 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8e900a41-c222-4bb7-b8da-5d038165be5b" (UID: "8e900a41-c222-4bb7-b8da-5d038165be5b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.598469 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-additional-scripts\") pod \"8e900a41-c222-4bb7-b8da-5d038165be5b\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.599198 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp6gq\" (UniqueName: \"kubernetes.io/projected/8e900a41-c222-4bb7-b8da-5d038165be5b-kube-api-access-zp6gq\") pod \"8e900a41-c222-4bb7-b8da-5d038165be5b\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.599272 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-scripts\") pod \"8e900a41-c222-4bb7-b8da-5d038165be5b\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.599304 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run\") pod \"8e900a41-c222-4bb7-b8da-5d038165be5b\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.599335 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-log-ovn\") pod \"8e900a41-c222-4bb7-b8da-5d038165be5b\" (UID: \"8e900a41-c222-4bb7-b8da-5d038165be5b\") " Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.599468 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "8e900a41-c222-4bb7-b8da-5d038165be5b" (UID: "8e900a41-c222-4bb7-b8da-5d038165be5b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.599763 4944 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.599778 4944 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.599803 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8e900a41-c222-4bb7-b8da-5d038165be5b" (UID: "8e900a41-c222-4bb7-b8da-5d038165be5b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.599820 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run" (OuterVolumeSpecName: "var-run") pod "8e900a41-c222-4bb7-b8da-5d038165be5b" (UID: "8e900a41-c222-4bb7-b8da-5d038165be5b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.600313 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-scripts" (OuterVolumeSpecName: "scripts") pod "8e900a41-c222-4bb7-b8da-5d038165be5b" (UID: "8e900a41-c222-4bb7-b8da-5d038165be5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.605945 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e900a41-c222-4bb7-b8da-5d038165be5b-kube-api-access-zp6gq" (OuterVolumeSpecName: "kube-api-access-zp6gq") pod "8e900a41-c222-4bb7-b8da-5d038165be5b" (UID: "8e900a41-c222-4bb7-b8da-5d038165be5b"). InnerVolumeSpecName "kube-api-access-zp6gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.703845 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp6gq\" (UniqueName: \"kubernetes.io/projected/8e900a41-c222-4bb7-b8da-5d038165be5b-kube-api-access-zp6gq\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.705085 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e900a41-c222-4bb7-b8da-5d038165be5b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.705098 4944 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:06 crc kubenswrapper[4944]: I1124 10:27:06.705112 4944 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e900a41-c222-4bb7-b8da-5d038165be5b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:07 crc kubenswrapper[4944]: I1124 10:27:07.020487 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-hsl6x"] Nov 24 10:27:07 crc kubenswrapper[4944]: W1124 10:27:07.039336 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47d6bd13_b381_4117_84a3_5418a76f2d24.slice/crio-eb2b476777c3385258c7a3849e508194cd7dbf733c08d1a020fd327d43c362a4 WatchSource:0}: Error finding container eb2b476777c3385258c7a3849e508194cd7dbf733c08d1a020fd327d43c362a4: Status 404 returned error can't find the container with id eb2b476777c3385258c7a3849e508194cd7dbf733c08d1a020fd327d43c362a4 Nov 24 10:27:07 crc kubenswrapper[4944]: I1124 10:27:07.082920 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dlb49-config-zx6fj" event={"ID":"8e900a41-c222-4bb7-b8da-5d038165be5b","Type":"ContainerDied","Data":"354bd92a08979d553389fc5021ea88727a5a754ce8d338bfd2d17892e3a87929"} Nov 24 10:27:07 crc kubenswrapper[4944]: I1124 10:27:07.082962 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="354bd92a08979d553389fc5021ea88727a5a754ce8d338bfd2d17892e3a87929" Nov 24 10:27:07 crc kubenswrapper[4944]: I1124 10:27:07.083007 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dlb49-config-zx6fj" Nov 24 10:27:07 crc kubenswrapper[4944]: I1124 10:27:07.085436 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-hsl6x" event={"ID":"47d6bd13-b381-4117-84a3-5418a76f2d24","Type":"ContainerStarted","Data":"eb2b476777c3385258c7a3849e508194cd7dbf733c08d1a020fd327d43c362a4"} Nov 24 10:27:07 crc kubenswrapper[4944]: I1124 10:27:07.277095 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:27:07 crc kubenswrapper[4944]: E1124 10:27:07.277555 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:27:07 crc kubenswrapper[4944]: I1124 10:27:07.532848 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-dlb49-config-zx6fj"] Nov 24 10:27:07 crc kubenswrapper[4944]: I1124 10:27:07.542426 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-dlb49-config-zx6fj"] Nov 24 10:27:08 crc kubenswrapper[4944]: I1124 10:27:08.095377 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-m8c9f" event={"ID":"36c49dd9-e55b-44f7-9dd6-50ec44c9fde2","Type":"ContainerStarted","Data":"dd6a78dedc6ab5671a3ebbd78f4509ff4a4825fc5b941db7b9000324e524dd4f"} Nov 24 10:27:08 crc kubenswrapper[4944]: I1124 10:27:08.096038 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:27:08 crc kubenswrapper[4944]: I1124 10:27:08.097083 4944 generic.go:334] "Generic (PLEG): container finished" podID="db61872a-0b2a-4fbb-b3ee-36570be61ff6" containerID="3411d8429ceda975ea7f346de4c10ed73bd99bd67ad1b8cb4af440a40f6dbf6c" exitCode=0 Nov 24 10:27:08 crc kubenswrapper[4944]: I1124 10:27:08.097163 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-g572j" event={"ID":"db61872a-0b2a-4fbb-b3ee-36570be61ff6","Type":"ContainerDied","Data":"3411d8429ceda975ea7f346de4c10ed73bd99bd67ad1b8cb4af440a40f6dbf6c"} Nov 24 10:27:08 crc kubenswrapper[4944]: I1124 10:27:08.099204 4944 generic.go:334] "Generic (PLEG): container finished" podID="47d6bd13-b381-4117-84a3-5418a76f2d24" containerID="ab8146ee72b37ba2c492d32d88f38d907be18b127b614a85fb669e060992a9d7" exitCode=0 Nov 24 10:27:08 crc kubenswrapper[4944]: I1124 10:27:08.099247 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-hsl6x" event={"ID":"47d6bd13-b381-4117-84a3-5418a76f2d24","Type":"ContainerDied","Data":"ab8146ee72b37ba2c492d32d88f38d907be18b127b614a85fb669e060992a9d7"} Nov 24 10:27:08 crc kubenswrapper[4944]: I1124 10:27:08.129261 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-m8c9f" podStartSLOduration=3.012574566 podStartE2EDuration="14.129242s" podCreationTimestamp="2025-11-24 10:26:54 +0000 UTC" firstStartedPulling="2025-11-24 10:26:55.807587287 +0000 UTC m=+5676.342027749" lastFinishedPulling="2025-11-24 10:27:06.924254721 +0000 UTC m=+5687.458695183" observedRunningTime="2025-11-24 10:27:08.11794705 +0000 UTC m=+5688.652387512" watchObservedRunningTime="2025-11-24 10:27:08.129242 +0000 UTC m=+5688.663682462" Nov 24 10:27:08 crc kubenswrapper[4944]: I1124 10:27:08.286851 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e900a41-c222-4bb7-b8da-5d038165be5b" path="/var/lib/kubelet/pods/8e900a41-c222-4bb7-b8da-5d038165be5b/volumes" Nov 24 10:27:09 crc kubenswrapper[4944]: I1124 10:27:09.110750 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-hsl6x" event={"ID":"47d6bd13-b381-4117-84a3-5418a76f2d24","Type":"ContainerStarted","Data":"0a5a72c6a6de28b80dace142d0f5de9523f8f98573e6ceb13cbf7eaaf2c48d82"} Nov 24 10:27:09 crc kubenswrapper[4944]: I1124 10:27:09.130949 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-hsl6x" podStartSLOduration=6.130930235 podStartE2EDuration="6.130930235s" podCreationTimestamp="2025-11-24 10:27:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:27:09.128384434 +0000 UTC m=+5689.662824906" watchObservedRunningTime="2025-11-24 10:27:09.130930235 +0000 UTC m=+5689.665370697" Nov 24 10:27:10 crc kubenswrapper[4944]: I1124 10:27:10.125386 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-g572j" event={"ID":"db61872a-0b2a-4fbb-b3ee-36570be61ff6","Type":"ContainerStarted","Data":"44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7"} Nov 24 10:27:10 crc kubenswrapper[4944]: I1124 10:27:10.169099 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-g572j" podStartSLOduration=2.203749134 podStartE2EDuration="15.169061331s" podCreationTimestamp="2025-11-24 10:26:55 +0000 UTC" firstStartedPulling="2025-11-24 10:26:56.463525763 +0000 UTC m=+5676.997966225" lastFinishedPulling="2025-11-24 10:27:09.42883796 +0000 UTC m=+5689.963278422" observedRunningTime="2025-11-24 10:27:10.154503137 +0000 UTC m=+5690.688943599" watchObservedRunningTime="2025-11-24 10:27:10.169061331 +0000 UTC m=+5690.703501793" Nov 24 10:27:13 crc kubenswrapper[4944]: I1124 10:27:13.150543 4944 generic.go:334] "Generic (PLEG): container finished" podID="47d6bd13-b381-4117-84a3-5418a76f2d24" containerID="0a5a72c6a6de28b80dace142d0f5de9523f8f98573e6ceb13cbf7eaaf2c48d82" exitCode=0 Nov 24 10:27:13 crc kubenswrapper[4944]: I1124 10:27:13.150613 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-hsl6x" event={"ID":"47d6bd13-b381-4117-84a3-5418a76f2d24","Type":"ContainerDied","Data":"0a5a72c6a6de28b80dace142d0f5de9523f8f98573e6ceb13cbf7eaaf2c48d82"} Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.492181 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.659780 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-scripts\") pod \"47d6bd13-b381-4117-84a3-5418a76f2d24\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.659916 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data\") pod \"47d6bd13-b381-4117-84a3-5418a76f2d24\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.659984 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-combined-ca-bundle\") pod \"47d6bd13-b381-4117-84a3-5418a76f2d24\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.660078 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data-merged\") pod \"47d6bd13-b381-4117-84a3-5418a76f2d24\" (UID: \"47d6bd13-b381-4117-84a3-5418a76f2d24\") " Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.671226 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data" (OuterVolumeSpecName: "config-data") pod "47d6bd13-b381-4117-84a3-5418a76f2d24" (UID: "47d6bd13-b381-4117-84a3-5418a76f2d24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.673917 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-scripts" (OuterVolumeSpecName: "scripts") pod "47d6bd13-b381-4117-84a3-5418a76f2d24" (UID: "47d6bd13-b381-4117-84a3-5418a76f2d24"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.685064 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "47d6bd13-b381-4117-84a3-5418a76f2d24" (UID: "47d6bd13-b381-4117-84a3-5418a76f2d24"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.686044 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47d6bd13-b381-4117-84a3-5418a76f2d24" (UID: "47d6bd13-b381-4117-84a3-5418a76f2d24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.762518 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.762579 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.762590 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d6bd13-b381-4117-84a3-5418a76f2d24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:14 crc kubenswrapper[4944]: I1124 10:27:14.762609 4944 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/47d6bd13-b381-4117-84a3-5418a76f2d24-config-data-merged\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:15 crc kubenswrapper[4944]: I1124 10:27:15.175167 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-hsl6x" event={"ID":"47d6bd13-b381-4117-84a3-5418a76f2d24","Type":"ContainerDied","Data":"eb2b476777c3385258c7a3849e508194cd7dbf733c08d1a020fd327d43c362a4"} Nov 24 10:27:15 crc kubenswrapper[4944]: I1124 10:27:15.175211 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb2b476777c3385258c7a3849e508194cd7dbf733c08d1a020fd327d43c362a4" Nov 24 10:27:15 crc kubenswrapper[4944]: I1124 10:27:15.175275 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-hsl6x" Nov 24 10:27:22 crc kubenswrapper[4944]: I1124 10:27:22.277327 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:27:22 crc kubenswrapper[4944]: E1124 10:27:22.278164 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:27:25 crc kubenswrapper[4944]: I1124 10:27:25.008036 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-m8c9f" Nov 24 10:27:31 crc kubenswrapper[4944]: I1124 10:27:31.285249 4944 scope.go:117] "RemoveContainer" containerID="cc8f9086bedc6d81ee3f7d8d6a18104f46373ed86032765f5f842f919418018b" Nov 24 10:27:31 crc kubenswrapper[4944]: I1124 10:27:31.325907 4944 scope.go:117] "RemoveContainer" containerID="26d2d171e762e827ca65a927fec0b123e36bd7469d4fa9ee92228d906ebef7f1" Nov 24 10:27:31 crc kubenswrapper[4944]: I1124 10:27:31.412299 4944 scope.go:117] "RemoveContainer" containerID="e7a1d58325c3cc36b5958c751173ec3e753c9be72c28474a0d8802843a24d0b4" Nov 24 10:27:31 crc kubenswrapper[4944]: I1124 10:27:31.450039 4944 scope.go:117] "RemoveContainer" containerID="e3f3e2983ce95602958fd55b450d5259f2022bf4446fd4ebb13af3b473aba9ac" Nov 24 10:27:33 crc kubenswrapper[4944]: I1124 10:27:33.188835 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-g572j"] Nov 24 10:27:33 crc kubenswrapper[4944]: I1124 10:27:33.189527 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-59f8cff499-g572j" podUID="db61872a-0b2a-4fbb-b3ee-36570be61ff6" containerName="octavia-amphora-httpd" containerID="cri-o://44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7" gracePeriod=30 Nov 24 10:27:33 crc kubenswrapper[4944]: I1124 10:27:33.789329 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-g572j" Nov 24 10:27:33 crc kubenswrapper[4944]: I1124 10:27:33.952999 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/db61872a-0b2a-4fbb-b3ee-36570be61ff6-httpd-config\") pod \"db61872a-0b2a-4fbb-b3ee-36570be61ff6\" (UID: \"db61872a-0b2a-4fbb-b3ee-36570be61ff6\") " Nov 24 10:27:33 crc kubenswrapper[4944]: I1124 10:27:33.953136 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/db61872a-0b2a-4fbb-b3ee-36570be61ff6-amphora-image\") pod \"db61872a-0b2a-4fbb-b3ee-36570be61ff6\" (UID: \"db61872a-0b2a-4fbb-b3ee-36570be61ff6\") " Nov 24 10:27:33 crc kubenswrapper[4944]: I1124 10:27:33.992580 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db61872a-0b2a-4fbb-b3ee-36570be61ff6-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "db61872a-0b2a-4fbb-b3ee-36570be61ff6" (UID: "db61872a-0b2a-4fbb-b3ee-36570be61ff6"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.051432 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db61872a-0b2a-4fbb-b3ee-36570be61ff6-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "db61872a-0b2a-4fbb-b3ee-36570be61ff6" (UID: "db61872a-0b2a-4fbb-b3ee-36570be61ff6"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.055508 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/db61872a-0b2a-4fbb-b3ee-36570be61ff6-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.055546 4944 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/db61872a-0b2a-4fbb-b3ee-36570be61ff6-amphora-image\") on node \"crc\" DevicePath \"\"" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.282893 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:27:34 crc kubenswrapper[4944]: E1124 10:27:34.284407 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.391702 4944 generic.go:334] "Generic (PLEG): container finished" podID="db61872a-0b2a-4fbb-b3ee-36570be61ff6" containerID="44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7" exitCode=0 Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.391756 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-g572j" event={"ID":"db61872a-0b2a-4fbb-b3ee-36570be61ff6","Type":"ContainerDied","Data":"44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7"} Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.392920 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-g572j" event={"ID":"db61872a-0b2a-4fbb-b3ee-36570be61ff6","Type":"ContainerDied","Data":"322ac47f69dbda1a456c16051db0524aed6519f33e9b7f036c3a0fbe59e297e1"} Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.392152 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-g572j" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.393032 4944 scope.go:117] "RemoveContainer" containerID="44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.454121 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-g572j"] Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.467480 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-g572j"] Nov 24 10:27:34 crc kubenswrapper[4944]: E1124 10:27:34.476296 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb61872a_0b2a_4fbb_b3ee_36570be61ff6.slice/crio-322ac47f69dbda1a456c16051db0524aed6519f33e9b7f036c3a0fbe59e297e1\": RecentStats: unable to find data in memory cache]" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.509618 4944 scope.go:117] "RemoveContainer" containerID="3411d8429ceda975ea7f346de4c10ed73bd99bd67ad1b8cb4af440a40f6dbf6c" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.528809 4944 scope.go:117] "RemoveContainer" containerID="44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7" Nov 24 10:27:34 crc kubenswrapper[4944]: E1124 10:27:34.531602 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7\": container with ID starting with 44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7 not found: ID does not exist" containerID="44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.531648 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7"} err="failed to get container status \"44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7\": rpc error: code = NotFound desc = could not find container \"44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7\": container with ID starting with 44ccafc6b8b320d4a7d85660f27963af6d92015a571a2216095766cd94e77ce7 not found: ID does not exist" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.531676 4944 scope.go:117] "RemoveContainer" containerID="3411d8429ceda975ea7f346de4c10ed73bd99bd67ad1b8cb4af440a40f6dbf6c" Nov 24 10:27:34 crc kubenswrapper[4944]: E1124 10:27:34.532297 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3411d8429ceda975ea7f346de4c10ed73bd99bd67ad1b8cb4af440a40f6dbf6c\": container with ID starting with 3411d8429ceda975ea7f346de4c10ed73bd99bd67ad1b8cb4af440a40f6dbf6c not found: ID does not exist" containerID="3411d8429ceda975ea7f346de4c10ed73bd99bd67ad1b8cb4af440a40f6dbf6c" Nov 24 10:27:34 crc kubenswrapper[4944]: I1124 10:27:34.532337 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3411d8429ceda975ea7f346de4c10ed73bd99bd67ad1b8cb4af440a40f6dbf6c"} err="failed to get container status \"3411d8429ceda975ea7f346de4c10ed73bd99bd67ad1b8cb4af440a40f6dbf6c\": rpc error: code = NotFound desc = could not find container \"3411d8429ceda975ea7f346de4c10ed73bd99bd67ad1b8cb4af440a40f6dbf6c\": container with ID starting with 3411d8429ceda975ea7f346de4c10ed73bd99bd67ad1b8cb4af440a40f6dbf6c not found: ID does not exist" Nov 24 10:27:36 crc kubenswrapper[4944]: I1124 10:27:36.294202 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db61872a-0b2a-4fbb-b3ee-36570be61ff6" path="/var/lib/kubelet/pods/db61872a-0b2a-4fbb-b3ee-36570be61ff6/volumes" Nov 24 10:27:48 crc kubenswrapper[4944]: I1124 10:27:48.277221 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:27:48 crc kubenswrapper[4944]: E1124 10:27:48.277928 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.755908 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-85x5v"] Nov 24 10:27:58 crc kubenswrapper[4944]: E1124 10:27:58.756931 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d6bd13-b381-4117-84a3-5418a76f2d24" containerName="octavia-db-sync" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.756950 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d6bd13-b381-4117-84a3-5418a76f2d24" containerName="octavia-db-sync" Nov 24 10:27:58 crc kubenswrapper[4944]: E1124 10:27:58.756964 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e900a41-c222-4bb7-b8da-5d038165be5b" containerName="ovn-config" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.756971 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e900a41-c222-4bb7-b8da-5d038165be5b" containerName="ovn-config" Nov 24 10:27:58 crc kubenswrapper[4944]: E1124 10:27:58.756997 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db61872a-0b2a-4fbb-b3ee-36570be61ff6" containerName="init" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.757006 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="db61872a-0b2a-4fbb-b3ee-36570be61ff6" containerName="init" Nov 24 10:27:58 crc kubenswrapper[4944]: E1124 10:27:58.757023 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d6bd13-b381-4117-84a3-5418a76f2d24" containerName="init" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.757030 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d6bd13-b381-4117-84a3-5418a76f2d24" containerName="init" Nov 24 10:27:58 crc kubenswrapper[4944]: E1124 10:27:58.757073 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db61872a-0b2a-4fbb-b3ee-36570be61ff6" containerName="octavia-amphora-httpd" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.757082 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="db61872a-0b2a-4fbb-b3ee-36570be61ff6" containerName="octavia-amphora-httpd" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.757296 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="db61872a-0b2a-4fbb-b3ee-36570be61ff6" containerName="octavia-amphora-httpd" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.757327 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e900a41-c222-4bb7-b8da-5d038165be5b" containerName="ovn-config" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.757349 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d6bd13-b381-4117-84a3-5418a76f2d24" containerName="octavia-db-sync" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.758764 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.762031 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.762073 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.762722 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.768922 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-85x5v"] Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.836559 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/80b0bc60-ac8c-440b-9323-1090e5121d0a-hm-ports\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.836607 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-scripts\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.836692 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-config-data\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.836766 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-combined-ca-bundle\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.836810 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/80b0bc60-ac8c-440b-9323-1090e5121d0a-config-data-merged\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.836839 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-amphora-certs\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.938722 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/80b0bc60-ac8c-440b-9323-1090e5121d0a-hm-ports\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.939108 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-scripts\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.939414 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-config-data\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.939687 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-combined-ca-bundle\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.939872 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/80b0bc60-ac8c-440b-9323-1090e5121d0a-config-data-merged\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.939985 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-amphora-certs\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.940301 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/80b0bc60-ac8c-440b-9323-1090e5121d0a-config-data-merged\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.940975 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/80b0bc60-ac8c-440b-9323-1090e5121d0a-hm-ports\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.944989 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-amphora-certs\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.945022 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-scripts\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.950782 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-combined-ca-bundle\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:58 crc kubenswrapper[4944]: I1124 10:27:58.952643 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b0bc60-ac8c-440b-9323-1090e5121d0a-config-data\") pod \"octavia-healthmanager-85x5v\" (UID: \"80b0bc60-ac8c-440b-9323-1090e5121d0a\") " pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.081647 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.276930 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:27:59 crc kubenswrapper[4944]: E1124 10:27:59.278197 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.705940 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-85x5v"] Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.784972 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-8mfrk"] Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.786584 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.789799 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.790114 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.793865 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-8mfrk"] Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.859911 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-combined-ca-bundle\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.860174 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-amphora-certs\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.860223 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-config-data\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.860504 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1352ff2a-e445-4c7c-930d-6356a19543eb-hm-ports\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.860554 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-scripts\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.860644 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1352ff2a-e445-4c7c-930d-6356a19543eb-config-data-merged\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.963564 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-combined-ca-bundle\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.963839 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-amphora-certs\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.963923 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-config-data\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.964067 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1352ff2a-e445-4c7c-930d-6356a19543eb-hm-ports\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.964102 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-scripts\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.964152 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1352ff2a-e445-4c7c-930d-6356a19543eb-config-data-merged\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.964579 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1352ff2a-e445-4c7c-930d-6356a19543eb-config-data-merged\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.965010 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1352ff2a-e445-4c7c-930d-6356a19543eb-hm-ports\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.970098 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-config-data\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.970646 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-scripts\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.970853 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-combined-ca-bundle\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:27:59 crc kubenswrapper[4944]: I1124 10:27:59.971322 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/1352ff2a-e445-4c7c-930d-6356a19543eb-amphora-certs\") pod \"octavia-housekeeping-8mfrk\" (UID: \"1352ff2a-e445-4c7c-930d-6356a19543eb\") " pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:28:00 crc kubenswrapper[4944]: I1124 10:28:00.112790 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:28:00 crc kubenswrapper[4944]: I1124 10:28:00.651516 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-85x5v" event={"ID":"80b0bc60-ac8c-440b-9323-1090e5121d0a","Type":"ContainerStarted","Data":"805cae69eac3cdc03607a36f382b8b2e5fb71e7dcd5fb655a0bb6e2aef7e59d6"} Nov 24 10:28:00 crc kubenswrapper[4944]: I1124 10:28:00.651657 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-85x5v" event={"ID":"80b0bc60-ac8c-440b-9323-1090e5121d0a","Type":"ContainerStarted","Data":"3db844a4b7e47fb273a1294bf41c79bba74221f668063908cdd2f7d651f3ce3d"} Nov 24 10:28:00 crc kubenswrapper[4944]: I1124 10:28:00.657410 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-8mfrk"] Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.635532 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-q5tb6"] Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.639164 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.643075 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.643259 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.646400 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-q5tb6"] Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.673320 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-8mfrk" event={"ID":"1352ff2a-e445-4c7c-930d-6356a19543eb","Type":"ContainerStarted","Data":"fc3aadb1bce833f2ba725705eb1e4a0a51fcc80f14dfe7e92afcb2543aa16e32"} Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.709132 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-amphora-certs\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.709189 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-scripts\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.709211 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/a22f171c-b444-4735-8c78-8ee9f41702a3-hm-ports\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.709248 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-combined-ca-bundle\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.709268 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-config-data\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.709364 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a22f171c-b444-4735-8c78-8ee9f41702a3-config-data-merged\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.811529 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-amphora-certs\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.811598 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-scripts\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.811627 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/a22f171c-b444-4735-8c78-8ee9f41702a3-hm-ports\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.811675 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-combined-ca-bundle\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.811699 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-config-data\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.811809 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a22f171c-b444-4735-8c78-8ee9f41702a3-config-data-merged\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.812600 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a22f171c-b444-4735-8c78-8ee9f41702a3-config-data-merged\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.814755 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/a22f171c-b444-4735-8c78-8ee9f41702a3-hm-ports\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.819347 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-amphora-certs\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.819637 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-config-data\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.819906 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-scripts\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.832672 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a22f171c-b444-4735-8c78-8ee9f41702a3-combined-ca-bundle\") pod \"octavia-worker-q5tb6\" (UID: \"a22f171c-b444-4735-8c78-8ee9f41702a3\") " pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:01 crc kubenswrapper[4944]: I1124 10:28:01.973670 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:02 crc kubenswrapper[4944]: I1124 10:28:02.570386 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-q5tb6"] Nov 24 10:28:02 crc kubenswrapper[4944]: W1124 10:28:02.580748 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda22f171c_b444_4735_8c78_8ee9f41702a3.slice/crio-9698f5d019c6ff668b364453eded160888ce8daedeb63cc3d2fbe54a895b2d24 WatchSource:0}: Error finding container 9698f5d019c6ff668b364453eded160888ce8daedeb63cc3d2fbe54a895b2d24: Status 404 returned error can't find the container with id 9698f5d019c6ff668b364453eded160888ce8daedeb63cc3d2fbe54a895b2d24 Nov 24 10:28:02 crc kubenswrapper[4944]: I1124 10:28:02.682676 4944 generic.go:334] "Generic (PLEG): container finished" podID="80b0bc60-ac8c-440b-9323-1090e5121d0a" containerID="805cae69eac3cdc03607a36f382b8b2e5fb71e7dcd5fb655a0bb6e2aef7e59d6" exitCode=0 Nov 24 10:28:02 crc kubenswrapper[4944]: I1124 10:28:02.682776 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-85x5v" event={"ID":"80b0bc60-ac8c-440b-9323-1090e5121d0a","Type":"ContainerDied","Data":"805cae69eac3cdc03607a36f382b8b2e5fb71e7dcd5fb655a0bb6e2aef7e59d6"} Nov 24 10:28:02 crc kubenswrapper[4944]: I1124 10:28:02.686272 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-q5tb6" event={"ID":"a22f171c-b444-4735-8c78-8ee9f41702a3","Type":"ContainerStarted","Data":"9698f5d019c6ff668b364453eded160888ce8daedeb63cc3d2fbe54a895b2d24"} Nov 24 10:28:02 crc kubenswrapper[4944]: I1124 10:28:02.688483 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-8mfrk" event={"ID":"1352ff2a-e445-4c7c-930d-6356a19543eb","Type":"ContainerStarted","Data":"687d0e9f1d7898569341833f893bf4fa43077f1e7382626e05e3c97269cbfc61"} Nov 24 10:28:03 crc kubenswrapper[4944]: I1124 10:28:03.701914 4944 generic.go:334] "Generic (PLEG): container finished" podID="1352ff2a-e445-4c7c-930d-6356a19543eb" containerID="687d0e9f1d7898569341833f893bf4fa43077f1e7382626e05e3c97269cbfc61" exitCode=0 Nov 24 10:28:03 crc kubenswrapper[4944]: I1124 10:28:03.702006 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-8mfrk" event={"ID":"1352ff2a-e445-4c7c-930d-6356a19543eb","Type":"ContainerDied","Data":"687d0e9f1d7898569341833f893bf4fa43077f1e7382626e05e3c97269cbfc61"} Nov 24 10:28:03 crc kubenswrapper[4944]: I1124 10:28:03.708773 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-85x5v" event={"ID":"80b0bc60-ac8c-440b-9323-1090e5121d0a","Type":"ContainerStarted","Data":"75a330cffc164126b84ddc5988088d929c9c1c37e53defb84235c6126f0e9d59"} Nov 24 10:28:03 crc kubenswrapper[4944]: I1124 10:28:03.709127 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:28:03 crc kubenswrapper[4944]: I1124 10:28:03.748073 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-85x5v" podStartSLOduration=5.748024709 podStartE2EDuration="5.748024709s" podCreationTimestamp="2025-11-24 10:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:28:03.742937627 +0000 UTC m=+5744.277378099" watchObservedRunningTime="2025-11-24 10:28:03.748024709 +0000 UTC m=+5744.282465171" Nov 24 10:28:04 crc kubenswrapper[4944]: I1124 10:28:04.724780 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-q5tb6" event={"ID":"a22f171c-b444-4735-8c78-8ee9f41702a3","Type":"ContainerStarted","Data":"3a5030fa804ca5d5ee6a7f1f3c31a068f3e1048bc080eeeb32e2db03314aabb3"} Nov 24 10:28:04 crc kubenswrapper[4944]: I1124 10:28:04.731657 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-8mfrk" event={"ID":"1352ff2a-e445-4c7c-930d-6356a19543eb","Type":"ContainerStarted","Data":"6b424634edd370d953485dc0daebc3dfbe4a1f5e6ecda649b497bd93d36e3fbb"} Nov 24 10:28:04 crc kubenswrapper[4944]: I1124 10:28:04.792558 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-8mfrk" podStartSLOduration=4.593087505 podStartE2EDuration="5.792532458s" podCreationTimestamp="2025-11-24 10:27:59 +0000 UTC" firstStartedPulling="2025-11-24 10:28:00.672719085 +0000 UTC m=+5741.207159547" lastFinishedPulling="2025-11-24 10:28:01.872164038 +0000 UTC m=+5742.406604500" observedRunningTime="2025-11-24 10:28:04.79043836 +0000 UTC m=+5745.324878842" watchObservedRunningTime="2025-11-24 10:28:04.792532458 +0000 UTC m=+5745.326972920" Nov 24 10:28:05 crc kubenswrapper[4944]: I1124 10:28:05.741333 4944 generic.go:334] "Generic (PLEG): container finished" podID="a22f171c-b444-4735-8c78-8ee9f41702a3" containerID="3a5030fa804ca5d5ee6a7f1f3c31a068f3e1048bc080eeeb32e2db03314aabb3" exitCode=0 Nov 24 10:28:05 crc kubenswrapper[4944]: I1124 10:28:05.741393 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-q5tb6" event={"ID":"a22f171c-b444-4735-8c78-8ee9f41702a3","Type":"ContainerDied","Data":"3a5030fa804ca5d5ee6a7f1f3c31a068f3e1048bc080eeeb32e2db03314aabb3"} Nov 24 10:28:05 crc kubenswrapper[4944]: I1124 10:28:05.742878 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:28:06 crc kubenswrapper[4944]: I1124 10:28:06.754630 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-q5tb6" event={"ID":"a22f171c-b444-4735-8c78-8ee9f41702a3","Type":"ContainerStarted","Data":"3abf47f7a4f64ede79e3d137911e2ed8fe0f4375cb639e7e9957ba518ded0c90"} Nov 24 10:28:06 crc kubenswrapper[4944]: I1124 10:28:06.781318 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-q5tb6" podStartSLOduration=4.46467104 podStartE2EDuration="5.781296703s" podCreationTimestamp="2025-11-24 10:28:01 +0000 UTC" firstStartedPulling="2025-11-24 10:28:02.586446252 +0000 UTC m=+5743.120886714" lastFinishedPulling="2025-11-24 10:28:03.903071915 +0000 UTC m=+5744.437512377" observedRunningTime="2025-11-24 10:28:06.775031733 +0000 UTC m=+5747.309472225" watchObservedRunningTime="2025-11-24 10:28:06.781296703 +0000 UTC m=+5747.315737175" Nov 24 10:28:07 crc kubenswrapper[4944]: I1124 10:28:07.766907 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:12 crc kubenswrapper[4944]: I1124 10:28:12.277528 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:28:12 crc kubenswrapper[4944]: E1124 10:28:12.278025 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:28:14 crc kubenswrapper[4944]: I1124 10:28:14.111284 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-85x5v" Nov 24 10:28:15 crc kubenswrapper[4944]: I1124 10:28:15.148620 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-8mfrk" Nov 24 10:28:17 crc kubenswrapper[4944]: I1124 10:28:17.003362 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-q5tb6" Nov 24 10:28:27 crc kubenswrapper[4944]: I1124 10:28:27.276909 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:28:27 crc kubenswrapper[4944]: E1124 10:28:27.277666 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:28:40 crc kubenswrapper[4944]: I1124 10:28:40.303879 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:28:40 crc kubenswrapper[4944]: E1124 10:28:40.304694 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:28:54 crc kubenswrapper[4944]: I1124 10:28:54.277185 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:28:54 crc kubenswrapper[4944]: E1124 10:28:54.278031 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.808800 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-76c9bc6d57-587xx"] Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.820413 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.826979 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.827182 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.827326 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.827413 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-m6d4h" Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.843778 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-76c9bc6d57-587xx"] Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.911492 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-scripts\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.913759 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3fcd7395-6aad-4766-bc81-2ff7609a859a-horizon-secret-key\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.913846 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fcd7395-6aad-4766-bc81-2ff7609a859a-logs\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.914003 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-config-data\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.914202 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwpwm\" (UniqueName: \"kubernetes.io/projected/3fcd7395-6aad-4766-bc81-2ff7609a859a-kube-api-access-kwpwm\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.958860 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.959160 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="57e038b8-73fd-4a98-8d1d-d06df5909909" containerName="glance-log" containerID="cri-o://56f50d48225e27d57f356dfdf23f611e6f870edabed7672e2657ecac0f0619d6" gracePeriod=30 Nov 24 10:28:56 crc kubenswrapper[4944]: I1124 10:28:56.959800 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="57e038b8-73fd-4a98-8d1d-d06df5909909" containerName="glance-httpd" containerID="cri-o://4550c9ae6e4446c0b8c01687c8c450d2082237a87c609f713ab0005d55cd8139" gracePeriod=30 Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.024167 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fcd7395-6aad-4766-bc81-2ff7609a859a-logs\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.024247 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-config-data\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.024309 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwpwm\" (UniqueName: \"kubernetes.io/projected/3fcd7395-6aad-4766-bc81-2ff7609a859a-kube-api-access-kwpwm\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.024346 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-scripts\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.025215 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-scripts\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.025291 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3fcd7395-6aad-4766-bc81-2ff7609a859a-horizon-secret-key\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.027040 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-config-data\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.027273 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fcd7395-6aad-4766-bc81-2ff7609a859a-logs\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.049782 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3fcd7395-6aad-4766-bc81-2ff7609a859a-horizon-secret-key\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.057323 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-69b6d48bbf-jwqrd"] Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.059466 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.071032 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwpwm\" (UniqueName: \"kubernetes.io/projected/3fcd7395-6aad-4766-bc81-2ff7609a859a-kube-api-access-kwpwm\") pod \"horizon-76c9bc6d57-587xx\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.074135 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-69b6d48bbf-jwqrd"] Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.126703 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/573d39f0-12fd-404e-8c47-cd39e153cbe2-logs\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.126784 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66x67\" (UniqueName: \"kubernetes.io/projected/573d39f0-12fd-404e-8c47-cd39e153cbe2-kube-api-access-66x67\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.126818 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-config-data\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.126915 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/573d39f0-12fd-404e-8c47-cd39e153cbe2-horizon-secret-key\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.126952 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-scripts\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.154190 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.155245 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.155424 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" containerName="glance-log" containerID="cri-o://b87f06c33d5fb6ea6d1cdf4f6a4978d6bddfebcfd760111f2e11997ada543bbe" gracePeriod=30 Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.155805 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" containerName="glance-httpd" containerID="cri-o://2dd1a505008d8acd50a78166cca4c0e762847d1f8ebcf1e93189444b348e6485" gracePeriod=30 Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.211906 4944 generic.go:334] "Generic (PLEG): container finished" podID="57e038b8-73fd-4a98-8d1d-d06df5909909" containerID="56f50d48225e27d57f356dfdf23f611e6f870edabed7672e2657ecac0f0619d6" exitCode=143 Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.211947 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"57e038b8-73fd-4a98-8d1d-d06df5909909","Type":"ContainerDied","Data":"56f50d48225e27d57f356dfdf23f611e6f870edabed7672e2657ecac0f0619d6"} Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.228132 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66x67\" (UniqueName: \"kubernetes.io/projected/573d39f0-12fd-404e-8c47-cd39e153cbe2-kube-api-access-66x67\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.228186 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-config-data\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.228257 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/573d39f0-12fd-404e-8c47-cd39e153cbe2-horizon-secret-key\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.228277 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-scripts\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.228359 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/573d39f0-12fd-404e-8c47-cd39e153cbe2-logs\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.228730 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/573d39f0-12fd-404e-8c47-cd39e153cbe2-logs\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.230313 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-config-data\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.231705 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-scripts\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.235579 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/573d39f0-12fd-404e-8c47-cd39e153cbe2-horizon-secret-key\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.244061 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66x67\" (UniqueName: \"kubernetes.io/projected/573d39f0-12fd-404e-8c47-cd39e153cbe2-kube-api-access-66x67\") pod \"horizon-69b6d48bbf-jwqrd\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.448257 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.543094 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-76c9bc6d57-587xx"] Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.569471 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-549c569ffc-sf7qz"] Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.571411 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.601390 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-549c569ffc-sf7qz"] Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.636289 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-horizon-secret-key\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.636351 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-logs\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.636439 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsbsd\" (UniqueName: \"kubernetes.io/projected/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-kube-api-access-lsbsd\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.636465 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-scripts\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.636490 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-config-data\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.690181 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-76c9bc6d57-587xx"] Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.742255 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsbsd\" (UniqueName: \"kubernetes.io/projected/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-kube-api-access-lsbsd\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.742326 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-scripts\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.742364 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-config-data\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.742456 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-horizon-secret-key\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.742505 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-logs\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.742999 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-logs\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.744544 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-config-data\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.748585 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-scripts\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.752987 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-horizon-secret-key\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.761192 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsbsd\" (UniqueName: \"kubernetes.io/projected/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-kube-api-access-lsbsd\") pod \"horizon-549c569ffc-sf7qz\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:57 crc kubenswrapper[4944]: I1124 10:28:57.959525 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:28:58 crc kubenswrapper[4944]: I1124 10:28:58.024668 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-69b6d48bbf-jwqrd"] Nov 24 10:28:58 crc kubenswrapper[4944]: W1124 10:28:58.027499 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod573d39f0_12fd_404e_8c47_cd39e153cbe2.slice/crio-e895723c9019b2dcd23c2dc12caea30b43bc01f8fe5bf63b865c04fc4aa4348f WatchSource:0}: Error finding container e895723c9019b2dcd23c2dc12caea30b43bc01f8fe5bf63b865c04fc4aa4348f: Status 404 returned error can't find the container with id e895723c9019b2dcd23c2dc12caea30b43bc01f8fe5bf63b865c04fc4aa4348f Nov 24 10:28:58 crc kubenswrapper[4944]: I1124 10:28:58.225631 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b6d48bbf-jwqrd" event={"ID":"573d39f0-12fd-404e-8c47-cd39e153cbe2","Type":"ContainerStarted","Data":"e895723c9019b2dcd23c2dc12caea30b43bc01f8fe5bf63b865c04fc4aa4348f"} Nov 24 10:28:58 crc kubenswrapper[4944]: I1124 10:28:58.226909 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c9bc6d57-587xx" event={"ID":"3fcd7395-6aad-4766-bc81-2ff7609a859a","Type":"ContainerStarted","Data":"5cf65a33e49a670b148a06d83136e190dd58754f3885e3a9bb42216344f5f780"} Nov 24 10:28:58 crc kubenswrapper[4944]: I1124 10:28:58.230063 4944 generic.go:334] "Generic (PLEG): container finished" podID="cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" containerID="b87f06c33d5fb6ea6d1cdf4f6a4978d6bddfebcfd760111f2e11997ada543bbe" exitCode=143 Nov 24 10:28:58 crc kubenswrapper[4944]: I1124 10:28:58.230113 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f","Type":"ContainerDied","Data":"b87f06c33d5fb6ea6d1cdf4f6a4978d6bddfebcfd760111f2e11997ada543bbe"} Nov 24 10:28:58 crc kubenswrapper[4944]: I1124 10:28:58.415033 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-549c569ffc-sf7qz"] Nov 24 10:28:59 crc kubenswrapper[4944]: I1124 10:28:59.251716 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549c569ffc-sf7qz" event={"ID":"cb28ab61-6a32-4fa9-986b-be0b8e69b56d","Type":"ContainerStarted","Data":"8b70bc6924c13cd91aef1cfb433a096481252871374b8baa265df0baa82e26f4"} Nov 24 10:29:00 crc kubenswrapper[4944]: I1124 10:29:00.260886 4944 generic.go:334] "Generic (PLEG): container finished" podID="57e038b8-73fd-4a98-8d1d-d06df5909909" containerID="4550c9ae6e4446c0b8c01687c8c450d2082237a87c609f713ab0005d55cd8139" exitCode=0 Nov 24 10:29:00 crc kubenswrapper[4944]: I1124 10:29:00.261066 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"57e038b8-73fd-4a98-8d1d-d06df5909909","Type":"ContainerDied","Data":"4550c9ae6e4446c0b8c01687c8c450d2082237a87c609f713ab0005d55cd8139"} Nov 24 10:29:01 crc kubenswrapper[4944]: I1124 10:29:01.274235 4944 generic.go:334] "Generic (PLEG): container finished" podID="cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" containerID="2dd1a505008d8acd50a78166cca4c0e762847d1f8ebcf1e93189444b348e6485" exitCode=0 Nov 24 10:29:01 crc kubenswrapper[4944]: I1124 10:29:01.274282 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f","Type":"ContainerDied","Data":"2dd1a505008d8acd50a78166cca4c0e762847d1f8ebcf1e93189444b348e6485"} Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.483475 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.576424 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-httpd-run\") pod \"57e038b8-73fd-4a98-8d1d-d06df5909909\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.576559 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-ceph\") pod \"57e038b8-73fd-4a98-8d1d-d06df5909909\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.576616 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltztr\" (UniqueName: \"kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-kube-api-access-ltztr\") pod \"57e038b8-73fd-4a98-8d1d-d06df5909909\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.576692 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-config-data\") pod \"57e038b8-73fd-4a98-8d1d-d06df5909909\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.576752 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-combined-ca-bundle\") pod \"57e038b8-73fd-4a98-8d1d-d06df5909909\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.576785 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-logs\") pod \"57e038b8-73fd-4a98-8d1d-d06df5909909\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.576804 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-scripts\") pod \"57e038b8-73fd-4a98-8d1d-d06df5909909\" (UID: \"57e038b8-73fd-4a98-8d1d-d06df5909909\") " Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.577533 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-logs" (OuterVolumeSpecName: "logs") pod "57e038b8-73fd-4a98-8d1d-d06df5909909" (UID: "57e038b8-73fd-4a98-8d1d-d06df5909909"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.579992 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "57e038b8-73fd-4a98-8d1d-d06df5909909" (UID: "57e038b8-73fd-4a98-8d1d-d06df5909909"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.580540 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-ceph" (OuterVolumeSpecName: "ceph") pod "57e038b8-73fd-4a98-8d1d-d06df5909909" (UID: "57e038b8-73fd-4a98-8d1d-d06df5909909"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.582840 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-kube-api-access-ltztr" (OuterVolumeSpecName: "kube-api-access-ltztr") pod "57e038b8-73fd-4a98-8d1d-d06df5909909" (UID: "57e038b8-73fd-4a98-8d1d-d06df5909909"). InnerVolumeSpecName "kube-api-access-ltztr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.585037 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-scripts" (OuterVolumeSpecName: "scripts") pod "57e038b8-73fd-4a98-8d1d-d06df5909909" (UID: "57e038b8-73fd-4a98-8d1d-d06df5909909"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.638306 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57e038b8-73fd-4a98-8d1d-d06df5909909" (UID: "57e038b8-73fd-4a98-8d1d-d06df5909909"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.679097 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.679130 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.679139 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltztr\" (UniqueName: \"kubernetes.io/projected/57e038b8-73fd-4a98-8d1d-d06df5909909-kube-api-access-ltztr\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.679148 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.679157 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57e038b8-73fd-4a98-8d1d-d06df5909909-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.679165 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.685180 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-config-data" (OuterVolumeSpecName: "config-data") pod "57e038b8-73fd-4a98-8d1d-d06df5909909" (UID: "57e038b8-73fd-4a98-8d1d-d06df5909909"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:29:04 crc kubenswrapper[4944]: I1124 10:29:04.780910 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57e038b8-73fd-4a98-8d1d-d06df5909909-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.277282 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:29:05 crc kubenswrapper[4944]: E1124 10:29:05.278031 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.318149 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549c569ffc-sf7qz" event={"ID":"cb28ab61-6a32-4fa9-986b-be0b8e69b56d","Type":"ContainerStarted","Data":"1155fa3580f692cb5172297de3ebfc0c787f3a83617bff8eeed96ab2d859150d"} Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.318204 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549c569ffc-sf7qz" event={"ID":"cb28ab61-6a32-4fa9-986b-be0b8e69b56d","Type":"ContainerStarted","Data":"2edd0863e2905254a5e16dea1e193dd7b8f6cc94e96cdb119d18b6c6d963bb41"} Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.329478 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c9bc6d57-587xx" event={"ID":"3fcd7395-6aad-4766-bc81-2ff7609a859a","Type":"ContainerStarted","Data":"fe7e6a6f9a02cfcbcc024e48757dffc8f122d24577674139ea22a69099470c66"} Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.329538 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c9bc6d57-587xx" event={"ID":"3fcd7395-6aad-4766-bc81-2ff7609a859a","Type":"ContainerStarted","Data":"8069ddfd4b1bb0600b68b04547dcd3c9f426247aeaca6e14c31d7cef1f59305c"} Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.329716 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-76c9bc6d57-587xx" podUID="3fcd7395-6aad-4766-bc81-2ff7609a859a" containerName="horizon-log" containerID="cri-o://8069ddfd4b1bb0600b68b04547dcd3c9f426247aeaca6e14c31d7cef1f59305c" gracePeriod=30 Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.329836 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-76c9bc6d57-587xx" podUID="3fcd7395-6aad-4766-bc81-2ff7609a859a" containerName="horizon" containerID="cri-o://fe7e6a6f9a02cfcbcc024e48757dffc8f122d24577674139ea22a69099470c66" gracePeriod=30 Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.333772 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.335669 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f","Type":"ContainerDied","Data":"eaaa45491b18af4a02799d3f201bc5a8d89ef6da12e0cf9f1ffff743305ab6ac"} Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.335718 4944 scope.go:117] "RemoveContainer" containerID="2dd1a505008d8acd50a78166cca4c0e762847d1f8ebcf1e93189444b348e6485" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.342221 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b6d48bbf-jwqrd" event={"ID":"573d39f0-12fd-404e-8c47-cd39e153cbe2","Type":"ContainerStarted","Data":"88243b62391b15a2c3a44a986dfe608b5457e5c130bbf60882723f986e18d386"} Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.342267 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b6d48bbf-jwqrd" event={"ID":"573d39f0-12fd-404e-8c47-cd39e153cbe2","Type":"ContainerStarted","Data":"5955542405b349aeb8d0ef4b9ddea8cd5b9be5b6644b129f0159d056b4b130ef"} Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.348006 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-549c569ffc-sf7qz" podStartSLOduration=2.607124256 podStartE2EDuration="8.3479766s" podCreationTimestamp="2025-11-24 10:28:57 +0000 UTC" firstStartedPulling="2025-11-24 10:28:58.424621794 +0000 UTC m=+5798.959062256" lastFinishedPulling="2025-11-24 10:29:04.165474138 +0000 UTC m=+5804.699914600" observedRunningTime="2025-11-24 10:29:05.340802882 +0000 UTC m=+5805.875243354" watchObservedRunningTime="2025-11-24 10:29:05.3479766 +0000 UTC m=+5805.882417062" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.359857 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"57e038b8-73fd-4a98-8d1d-d06df5909909","Type":"ContainerDied","Data":"d66ffe6d68c243de044f8f352e1b7f1e90fcf90fbb837051f10aa96f82349c51"} Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.359989 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.379295 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-76c9bc6d57-587xx" podStartSLOduration=2.926445048 podStartE2EDuration="9.379274317s" podCreationTimestamp="2025-11-24 10:28:56 +0000 UTC" firstStartedPulling="2025-11-24 10:28:57.695656978 +0000 UTC m=+5798.230097440" lastFinishedPulling="2025-11-24 10:29:04.148486247 +0000 UTC m=+5804.682926709" observedRunningTime="2025-11-24 10:29:05.369866138 +0000 UTC m=+5805.904306600" watchObservedRunningTime="2025-11-24 10:29:05.379274317 +0000 UTC m=+5805.913714789" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.381244 4944 scope.go:117] "RemoveContainer" containerID="b87f06c33d5fb6ea6d1cdf4f6a4978d6bddfebcfd760111f2e11997ada543bbe" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.413516 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5vmk\" (UniqueName: \"kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-kube-api-access-c5vmk\") pod \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.434993 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-httpd-run\") pod \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.435109 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-logs\") pod \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.435186 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-config-data\") pod \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.435296 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-scripts\") pod \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.435348 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-ceph\") pod \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.435390 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-combined-ca-bundle\") pod \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\" (UID: \"cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f\") " Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.432387 4944 scope.go:117] "RemoveContainer" containerID="4550c9ae6e4446c0b8c01687c8c450d2082237a87c609f713ab0005d55cd8139" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.432573 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-kube-api-access-c5vmk" (OuterVolumeSpecName: "kube-api-access-c5vmk") pod "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" (UID: "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f"). InnerVolumeSpecName "kube-api-access-c5vmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.439644 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" (UID: "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.443133 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-logs" (OuterVolumeSpecName: "logs") pod "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" (UID: "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.450132 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.453313 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-ceph" (OuterVolumeSpecName: "ceph") pod "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" (UID: "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.472310 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-scripts" (OuterVolumeSpecName: "scripts") pod "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" (UID: "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.476496 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.489196 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:29:05 crc kubenswrapper[4944]: E1124 10:29:05.489760 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" containerName="glance-log" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.489777 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" containerName="glance-log" Nov 24 10:29:05 crc kubenswrapper[4944]: E1124 10:29:05.489802 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e038b8-73fd-4a98-8d1d-d06df5909909" containerName="glance-log" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.489809 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e038b8-73fd-4a98-8d1d-d06df5909909" containerName="glance-log" Nov 24 10:29:05 crc kubenswrapper[4944]: E1124 10:29:05.489820 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e038b8-73fd-4a98-8d1d-d06df5909909" containerName="glance-httpd" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.489828 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e038b8-73fd-4a98-8d1d-d06df5909909" containerName="glance-httpd" Nov 24 10:29:05 crc kubenswrapper[4944]: E1124 10:29:05.489860 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" containerName="glance-httpd" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.489868 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" containerName="glance-httpd" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.490121 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e038b8-73fd-4a98-8d1d-d06df5909909" containerName="glance-log" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.490143 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e038b8-73fd-4a98-8d1d-d06df5909909" containerName="glance-httpd" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.490154 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" containerName="glance-log" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.490167 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" containerName="glance-httpd" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.511014 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.513060 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-69b6d48bbf-jwqrd" podStartSLOduration=2.406024451 podStartE2EDuration="8.513025657s" podCreationTimestamp="2025-11-24 10:28:57 +0000 UTC" firstStartedPulling="2025-11-24 10:28:58.030600985 +0000 UTC m=+5798.565041447" lastFinishedPulling="2025-11-24 10:29:04.137602191 +0000 UTC m=+5804.672042653" observedRunningTime="2025-11-24 10:29:05.470788202 +0000 UTC m=+5806.005228694" watchObservedRunningTime="2025-11-24 10:29:05.513025657 +0000 UTC m=+5806.047466139" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.515650 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.516513 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-config-data" (OuterVolumeSpecName: "config-data") pod "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" (UID: "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.522611 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" (UID: "cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.538692 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.538734 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5vmk\" (UniqueName: \"kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-kube-api-access-c5vmk\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.538752 4944 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.538764 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.538779 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.538791 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.538801 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.551798 4944 scope.go:117] "RemoveContainer" containerID="56f50d48225e27d57f356dfdf23f611e6f870edabed7672e2657ecac0f0619d6" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.553708 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.640599 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/942ae185-df7d-4585-995a-dc253806d0fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.640647 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/942ae185-df7d-4585-995a-dc253806d0fe-logs\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.640705 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942ae185-df7d-4585-995a-dc253806d0fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.640870 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942ae185-df7d-4585-995a-dc253806d0fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.640994 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smb2k\" (UniqueName: \"kubernetes.io/projected/942ae185-df7d-4585-995a-dc253806d0fe-kube-api-access-smb2k\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.641021 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/942ae185-df7d-4585-995a-dc253806d0fe-ceph\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.641073 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942ae185-df7d-4585-995a-dc253806d0fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.742927 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/942ae185-df7d-4585-995a-dc253806d0fe-ceph\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.743239 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942ae185-df7d-4585-995a-dc253806d0fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.743348 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/942ae185-df7d-4585-995a-dc253806d0fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.743423 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/942ae185-df7d-4585-995a-dc253806d0fe-logs\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.743546 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942ae185-df7d-4585-995a-dc253806d0fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.743681 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942ae185-df7d-4585-995a-dc253806d0fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.743830 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smb2k\" (UniqueName: \"kubernetes.io/projected/942ae185-df7d-4585-995a-dc253806d0fe-kube-api-access-smb2k\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.743943 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/942ae185-df7d-4585-995a-dc253806d0fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.744370 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/942ae185-df7d-4585-995a-dc253806d0fe-logs\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.748208 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/942ae185-df7d-4585-995a-dc253806d0fe-ceph\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.748232 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942ae185-df7d-4585-995a-dc253806d0fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.748289 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942ae185-df7d-4585-995a-dc253806d0fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.748712 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942ae185-df7d-4585-995a-dc253806d0fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.759741 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smb2k\" (UniqueName: \"kubernetes.io/projected/942ae185-df7d-4585-995a-dc253806d0fe-kube-api-access-smb2k\") pod \"glance-default-external-api-0\" (UID: \"942ae185-df7d-4585-995a-dc253806d0fe\") " pod="openstack/glance-default-external-api-0" Nov 24 10:29:05 crc kubenswrapper[4944]: I1124 10:29:05.841975 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.291819 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57e038b8-73fd-4a98-8d1d-d06df5909909" path="/var/lib/kubelet/pods/57e038b8-73fd-4a98-8d1d-d06df5909909/volumes" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.381505 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.417158 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.445590 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.465492 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.465549 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.467289 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.474073 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.478522 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.572195 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.572264 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.572357 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.572826 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrzlx\" (UniqueName: \"kubernetes.io/projected/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-kube-api-access-qrzlx\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.572895 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-ceph\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.573498 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-logs\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.573548 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.675096 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrzlx\" (UniqueName: \"kubernetes.io/projected/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-kube-api-access-qrzlx\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.675162 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-ceph\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.675229 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-logs\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.675266 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.675301 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.675336 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.675355 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.675662 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-logs\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.675769 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.682972 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.690402 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-ceph\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.691001 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.694705 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.698823 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrzlx\" (UniqueName: \"kubernetes.io/projected/d2765b02-6d6d-42c2-bbf1-92ae32ef6662-kube-api-access-qrzlx\") pod \"glance-default-internal-api-0\" (UID: \"d2765b02-6d6d-42c2-bbf1-92ae32ef6662\") " pod="openstack/glance-default-internal-api-0" Nov 24 10:29:06 crc kubenswrapper[4944]: I1124 10:29:06.819380 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:07 crc kubenswrapper[4944]: I1124 10:29:07.154950 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:29:07 crc kubenswrapper[4944]: I1124 10:29:07.412772 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"942ae185-df7d-4585-995a-dc253806d0fe","Type":"ContainerStarted","Data":"e63bba776f806e537a95cc7de80eb386ce2e4da30cc76fe9e49edb2a9b9ac511"} Nov 24 10:29:07 crc kubenswrapper[4944]: I1124 10:29:07.412818 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"942ae185-df7d-4585-995a-dc253806d0fe","Type":"ContainerStarted","Data":"cdd786787aa113b551ea9c24ab79534920ff869b515bfcec6df5d12794fae4c5"} Nov 24 10:29:07 crc kubenswrapper[4944]: I1124 10:29:07.414334 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 10:29:07 crc kubenswrapper[4944]: W1124 10:29:07.421174 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2765b02_6d6d_42c2_bbf1_92ae32ef6662.slice/crio-86ca22b540648c3bb85c3e7ab840dd8c8edd8bfd52f373d8ea6f38b434f6515a WatchSource:0}: Error finding container 86ca22b540648c3bb85c3e7ab840dd8c8edd8bfd52f373d8ea6f38b434f6515a: Status 404 returned error can't find the container with id 86ca22b540648c3bb85c3e7ab840dd8c8edd8bfd52f373d8ea6f38b434f6515a Nov 24 10:29:07 crc kubenswrapper[4944]: I1124 10:29:07.448699 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:29:07 crc kubenswrapper[4944]: I1124 10:29:07.449198 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:29:07 crc kubenswrapper[4944]: I1124 10:29:07.960538 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:29:07 crc kubenswrapper[4944]: I1124 10:29:07.960963 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:29:08 crc kubenswrapper[4944]: I1124 10:29:08.292306 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f" path="/var/lib/kubelet/pods/cb4f3bda-48f5-4f1b-9a12-84f6b8c67e0f/volumes" Nov 24 10:29:08 crc kubenswrapper[4944]: I1124 10:29:08.440315 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"942ae185-df7d-4585-995a-dc253806d0fe","Type":"ContainerStarted","Data":"4e4077a13e6b95789ebdf4349ba8c0972e531c00329d2750d07634e16561d27b"} Nov 24 10:29:08 crc kubenswrapper[4944]: I1124 10:29:08.443737 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d2765b02-6d6d-42c2-bbf1-92ae32ef6662","Type":"ContainerStarted","Data":"c5d70f4c6c1321a764a4504a1304f3e9278ea0adfc5283db3d58a17cb2a7e11a"} Nov 24 10:29:08 crc kubenswrapper[4944]: I1124 10:29:08.443776 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d2765b02-6d6d-42c2-bbf1-92ae32ef6662","Type":"ContainerStarted","Data":"86ca22b540648c3bb85c3e7ab840dd8c8edd8bfd52f373d8ea6f38b434f6515a"} Nov 24 10:29:08 crc kubenswrapper[4944]: I1124 10:29:08.473115 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.473096124 podStartE2EDuration="3.473096124s" podCreationTimestamp="2025-11-24 10:29:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:29:08.470998437 +0000 UTC m=+5809.005438899" watchObservedRunningTime="2025-11-24 10:29:08.473096124 +0000 UTC m=+5809.007536596" Nov 24 10:29:09 crc kubenswrapper[4944]: I1124 10:29:09.038149 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-lnchq"] Nov 24 10:29:09 crc kubenswrapper[4944]: I1124 10:29:09.045845 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-24fb-account-create-5lptb"] Nov 24 10:29:09 crc kubenswrapper[4944]: I1124 10:29:09.057137 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-lnchq"] Nov 24 10:29:09 crc kubenswrapper[4944]: I1124 10:29:09.067550 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-24fb-account-create-5lptb"] Nov 24 10:29:09 crc kubenswrapper[4944]: I1124 10:29:09.453120 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d2765b02-6d6d-42c2-bbf1-92ae32ef6662","Type":"ContainerStarted","Data":"166e2be86f247cbeb22e75540a367cea6dbc0a0a01841cc7186d005693f5d6e4"} Nov 24 10:29:09 crc kubenswrapper[4944]: I1124 10:29:09.474338 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.474312362 podStartE2EDuration="3.474312362s" podCreationTimestamp="2025-11-24 10:29:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:29:09.46764624 +0000 UTC m=+5810.002086712" watchObservedRunningTime="2025-11-24 10:29:09.474312362 +0000 UTC m=+5810.008752824" Nov 24 10:29:10 crc kubenswrapper[4944]: I1124 10:29:10.310439 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ff6aaed-4959-4839-a354-26c2056dda5e" path="/var/lib/kubelet/pods/3ff6aaed-4959-4839-a354-26c2056dda5e/volumes" Nov 24 10:29:10 crc kubenswrapper[4944]: I1124 10:29:10.312311 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8ee2a60-9953-43cd-8c12-fb7c9e36c85a" path="/var/lib/kubelet/pods/d8ee2a60-9953-43cd-8c12-fb7c9e36c85a/volumes" Nov 24 10:29:14 crc kubenswrapper[4944]: I1124 10:29:14.027219 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-qrs6l"] Nov 24 10:29:14 crc kubenswrapper[4944]: I1124 10:29:14.036322 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-qrs6l"] Nov 24 10:29:14 crc kubenswrapper[4944]: I1124 10:29:14.288874 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="697fd7a2-cd31-45a5-8e12-3b05719a20c0" path="/var/lib/kubelet/pods/697fd7a2-cd31-45a5-8e12-3b05719a20c0/volumes" Nov 24 10:29:15 crc kubenswrapper[4944]: I1124 10:29:15.843122 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 10:29:15 crc kubenswrapper[4944]: I1124 10:29:15.843185 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 10:29:15 crc kubenswrapper[4944]: I1124 10:29:15.875909 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 10:29:15 crc kubenswrapper[4944]: I1124 10:29:15.887603 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 10:29:16 crc kubenswrapper[4944]: I1124 10:29:16.516170 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 10:29:16 crc kubenswrapper[4944]: I1124 10:29:16.516511 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 10:29:16 crc kubenswrapper[4944]: I1124 10:29:16.819980 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:16 crc kubenswrapper[4944]: I1124 10:29:16.820121 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:16 crc kubenswrapper[4944]: I1124 10:29:16.865736 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:16 crc kubenswrapper[4944]: I1124 10:29:16.865824 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:17 crc kubenswrapper[4944]: I1124 10:29:17.450572 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-69b6d48bbf-jwqrd" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Nov 24 10:29:17 crc kubenswrapper[4944]: I1124 10:29:17.524964 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:17 crc kubenswrapper[4944]: I1124 10:29:17.525655 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:17 crc kubenswrapper[4944]: I1124 10:29:17.961457 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-549c569ffc-sf7qz" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.108:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.108:8080: connect: connection refused" Nov 24 10:29:18 crc kubenswrapper[4944]: I1124 10:29:18.534521 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 10:29:18 crc kubenswrapper[4944]: I1124 10:29:18.534900 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 10:29:18 crc kubenswrapper[4944]: I1124 10:29:18.548162 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 10:29:18 crc kubenswrapper[4944]: I1124 10:29:18.555774 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 10:29:19 crc kubenswrapper[4944]: I1124 10:29:19.543550 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 10:29:19 crc kubenswrapper[4944]: I1124 10:29:19.543575 4944 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 10:29:19 crc kubenswrapper[4944]: I1124 10:29:19.621788 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:19 crc kubenswrapper[4944]: I1124 10:29:19.623959 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 10:29:20 crc kubenswrapper[4944]: I1124 10:29:20.285854 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:29:20 crc kubenswrapper[4944]: E1124 10:29:20.286478 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:29:29 crc kubenswrapper[4944]: I1124 10:29:29.358209 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:29:29 crc kubenswrapper[4944]: I1124 10:29:29.757893 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:29:31 crc kubenswrapper[4944]: I1124 10:29:31.081606 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:29:31 crc kubenswrapper[4944]: I1124 10:29:31.520994 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:29:31 crc kubenswrapper[4944]: I1124 10:29:31.586634 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69b6d48bbf-jwqrd"] Nov 24 10:29:31 crc kubenswrapper[4944]: I1124 10:29:31.627349 4944 scope.go:117] "RemoveContainer" containerID="fbb488a6adfa0c34cce9196736c83a62353ec6209398f2fea648fb462c89a43b" Nov 24 10:29:31 crc kubenswrapper[4944]: I1124 10:29:31.650322 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69b6d48bbf-jwqrd" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon-log" containerID="cri-o://5955542405b349aeb8d0ef4b9ddea8cd5b9be5b6644b129f0159d056b4b130ef" gracePeriod=30 Nov 24 10:29:31 crc kubenswrapper[4944]: I1124 10:29:31.650685 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69b6d48bbf-jwqrd" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon" containerID="cri-o://88243b62391b15a2c3a44a986dfe608b5457e5c130bbf60882723f986e18d386" gracePeriod=30 Nov 24 10:29:31 crc kubenswrapper[4944]: I1124 10:29:31.669277 4944 scope.go:117] "RemoveContainer" containerID="ac67d78c4a324f1429afe15211ee04394b93fb06e5d16567d56f7f1d01839387" Nov 24 10:29:31 crc kubenswrapper[4944]: I1124 10:29:31.702279 4944 scope.go:117] "RemoveContainer" containerID="d0778072724d69111fcee9fa5119a3aa735f4eed5a48bd890eb400cb5301487f" Nov 24 10:29:31 crc kubenswrapper[4944]: I1124 10:29:31.747219 4944 scope.go:117] "RemoveContainer" containerID="c1a160deae388ec1c7a13572c6b785e752aa15edebd970415346b103e75f1681" Nov 24 10:29:31 crc kubenswrapper[4944]: I1124 10:29:31.796329 4944 scope.go:117] "RemoveContainer" containerID="a79308575fa58b1f06e8a3c2da13f2e44ce9fa9b5031db96d89f6ca9517c741f" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.277406 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:29:35 crc kubenswrapper[4944]: E1124 10:29:35.278217 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.682788 4944 generic.go:334] "Generic (PLEG): container finished" podID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerID="88243b62391b15a2c3a44a986dfe608b5457e5c130bbf60882723f986e18d386" exitCode=0 Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.682880 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b6d48bbf-jwqrd" event={"ID":"573d39f0-12fd-404e-8c47-cd39e153cbe2","Type":"ContainerDied","Data":"88243b62391b15a2c3a44a986dfe608b5457e5c130bbf60882723f986e18d386"} Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.685259 4944 generic.go:334] "Generic (PLEG): container finished" podID="3fcd7395-6aad-4766-bc81-2ff7609a859a" containerID="fe7e6a6f9a02cfcbcc024e48757dffc8f122d24577674139ea22a69099470c66" exitCode=137 Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.685275 4944 generic.go:334] "Generic (PLEG): container finished" podID="3fcd7395-6aad-4766-bc81-2ff7609a859a" containerID="8069ddfd4b1bb0600b68b04547dcd3c9f426247aeaca6e14c31d7cef1f59305c" exitCode=137 Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.685282 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c9bc6d57-587xx" event={"ID":"3fcd7395-6aad-4766-bc81-2ff7609a859a","Type":"ContainerDied","Data":"fe7e6a6f9a02cfcbcc024e48757dffc8f122d24577674139ea22a69099470c66"} Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.685327 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c9bc6d57-587xx" event={"ID":"3fcd7395-6aad-4766-bc81-2ff7609a859a","Type":"ContainerDied","Data":"8069ddfd4b1bb0600b68b04547dcd3c9f426247aeaca6e14c31d7cef1f59305c"} Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.685342 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76c9bc6d57-587xx" event={"ID":"3fcd7395-6aad-4766-bc81-2ff7609a859a","Type":"ContainerDied","Data":"5cf65a33e49a670b148a06d83136e190dd58754f3885e3a9bb42216344f5f780"} Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.685355 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cf65a33e49a670b148a06d83136e190dd58754f3885e3a9bb42216344f5f780" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.746157 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.888692 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fcd7395-6aad-4766-bc81-2ff7609a859a-logs\") pod \"3fcd7395-6aad-4766-bc81-2ff7609a859a\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.888742 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwpwm\" (UniqueName: \"kubernetes.io/projected/3fcd7395-6aad-4766-bc81-2ff7609a859a-kube-api-access-kwpwm\") pod \"3fcd7395-6aad-4766-bc81-2ff7609a859a\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.888794 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-scripts\") pod \"3fcd7395-6aad-4766-bc81-2ff7609a859a\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.888941 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-config-data\") pod \"3fcd7395-6aad-4766-bc81-2ff7609a859a\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.889024 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3fcd7395-6aad-4766-bc81-2ff7609a859a-horizon-secret-key\") pod \"3fcd7395-6aad-4766-bc81-2ff7609a859a\" (UID: \"3fcd7395-6aad-4766-bc81-2ff7609a859a\") " Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.889180 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fcd7395-6aad-4766-bc81-2ff7609a859a-logs" (OuterVolumeSpecName: "logs") pod "3fcd7395-6aad-4766-bc81-2ff7609a859a" (UID: "3fcd7395-6aad-4766-bc81-2ff7609a859a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.889582 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fcd7395-6aad-4766-bc81-2ff7609a859a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.896300 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fcd7395-6aad-4766-bc81-2ff7609a859a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3fcd7395-6aad-4766-bc81-2ff7609a859a" (UID: "3fcd7395-6aad-4766-bc81-2ff7609a859a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.896336 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fcd7395-6aad-4766-bc81-2ff7609a859a-kube-api-access-kwpwm" (OuterVolumeSpecName: "kube-api-access-kwpwm") pod "3fcd7395-6aad-4766-bc81-2ff7609a859a" (UID: "3fcd7395-6aad-4766-bc81-2ff7609a859a"). InnerVolumeSpecName "kube-api-access-kwpwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.915298 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-config-data" (OuterVolumeSpecName: "config-data") pod "3fcd7395-6aad-4766-bc81-2ff7609a859a" (UID: "3fcd7395-6aad-4766-bc81-2ff7609a859a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.915992 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-scripts" (OuterVolumeSpecName: "scripts") pod "3fcd7395-6aad-4766-bc81-2ff7609a859a" (UID: "3fcd7395-6aad-4766-bc81-2ff7609a859a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.991258 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.991288 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3fcd7395-6aad-4766-bc81-2ff7609a859a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.991299 4944 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3fcd7395-6aad-4766-bc81-2ff7609a859a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:35 crc kubenswrapper[4944]: I1124 10:29:35.991312 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwpwm\" (UniqueName: \"kubernetes.io/projected/3fcd7395-6aad-4766-bc81-2ff7609a859a-kube-api-access-kwpwm\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:36 crc kubenswrapper[4944]: I1124 10:29:36.693382 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76c9bc6d57-587xx" Nov 24 10:29:36 crc kubenswrapper[4944]: I1124 10:29:36.715832 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-76c9bc6d57-587xx"] Nov 24 10:29:36 crc kubenswrapper[4944]: I1124 10:29:36.724322 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-76c9bc6d57-587xx"] Nov 24 10:29:37 crc kubenswrapper[4944]: I1124 10:29:37.450169 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-69b6d48bbf-jwqrd" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.288274 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fcd7395-6aad-4766-bc81-2ff7609a859a" path="/var/lib/kubelet/pods/3fcd7395-6aad-4766-bc81-2ff7609a859a/volumes" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.839340 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74876d679c-sbxz5"] Nov 24 10:29:38 crc kubenswrapper[4944]: E1124 10:29:38.839720 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fcd7395-6aad-4766-bc81-2ff7609a859a" containerName="horizon" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.839734 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fcd7395-6aad-4766-bc81-2ff7609a859a" containerName="horizon" Nov 24 10:29:38 crc kubenswrapper[4944]: E1124 10:29:38.839769 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fcd7395-6aad-4766-bc81-2ff7609a859a" containerName="horizon-log" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.839777 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fcd7395-6aad-4766-bc81-2ff7609a859a" containerName="horizon-log" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.839958 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fcd7395-6aad-4766-bc81-2ff7609a859a" containerName="horizon-log" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.839979 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fcd7395-6aad-4766-bc81-2ff7609a859a" containerName="horizon" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.840993 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.853271 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74876d679c-sbxz5"] Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.959660 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/946e72b9-d3e9-4823-8b87-2049e4a107ac-config-data\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.959724 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/946e72b9-d3e9-4823-8b87-2049e4a107ac-horizon-secret-key\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.959751 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/946e72b9-d3e9-4823-8b87-2049e4a107ac-scripts\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.959957 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/946e72b9-d3e9-4823-8b87-2049e4a107ac-logs\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:38 crc kubenswrapper[4944]: I1124 10:29:38.960176 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87ckd\" (UniqueName: \"kubernetes.io/projected/946e72b9-d3e9-4823-8b87-2049e4a107ac-kube-api-access-87ckd\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.062511 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/946e72b9-d3e9-4823-8b87-2049e4a107ac-config-data\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.062583 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/946e72b9-d3e9-4823-8b87-2049e4a107ac-horizon-secret-key\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.062609 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/946e72b9-d3e9-4823-8b87-2049e4a107ac-scripts\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.062684 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/946e72b9-d3e9-4823-8b87-2049e4a107ac-logs\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.062779 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87ckd\" (UniqueName: \"kubernetes.io/projected/946e72b9-d3e9-4823-8b87-2049e4a107ac-kube-api-access-87ckd\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.063376 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/946e72b9-d3e9-4823-8b87-2049e4a107ac-logs\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.063622 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/946e72b9-d3e9-4823-8b87-2049e4a107ac-scripts\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.064090 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/946e72b9-d3e9-4823-8b87-2049e4a107ac-config-data\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.076877 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/946e72b9-d3e9-4823-8b87-2049e4a107ac-horizon-secret-key\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.084693 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87ckd\" (UniqueName: \"kubernetes.io/projected/946e72b9-d3e9-4823-8b87-2049e4a107ac-kube-api-access-87ckd\") pod \"horizon-74876d679c-sbxz5\" (UID: \"946e72b9-d3e9-4823-8b87-2049e4a107ac\") " pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.160415 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.642331 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74876d679c-sbxz5"] Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.739454 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74876d679c-sbxz5" event={"ID":"946e72b9-d3e9-4823-8b87-2049e4a107ac","Type":"ContainerStarted","Data":"d89d2f9dd556eba209af070ba4ef594420f175fbae0ab47780fe80ccb976bb7d"} Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.971248 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-9l7mt"] Nov 24 10:29:39 crc kubenswrapper[4944]: I1124 10:29:39.972426 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-9l7mt" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.030174 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-9l7mt"] Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.082209 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95067a49-0792-4ee9-a0fd-6cec20b39955-operator-scripts\") pod \"heat-db-create-9l7mt\" (UID: \"95067a49-0792-4ee9-a0fd-6cec20b39955\") " pod="openstack/heat-db-create-9l7mt" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.082463 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd9ss\" (UniqueName: \"kubernetes.io/projected/95067a49-0792-4ee9-a0fd-6cec20b39955-kube-api-access-xd9ss\") pod \"heat-db-create-9l7mt\" (UID: \"95067a49-0792-4ee9-a0fd-6cec20b39955\") " pod="openstack/heat-db-create-9l7mt" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.084573 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-eaba-account-create-wdpst"] Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.086427 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-eaba-account-create-wdpst" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.093496 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.111123 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-eaba-account-create-wdpst"] Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.184593 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbca7358-993f-4558-b9b3-e8f41abfbcdf-operator-scripts\") pod \"heat-eaba-account-create-wdpst\" (UID: \"dbca7358-993f-4558-b9b3-e8f41abfbcdf\") " pod="openstack/heat-eaba-account-create-wdpst" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.184657 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95067a49-0792-4ee9-a0fd-6cec20b39955-operator-scripts\") pod \"heat-db-create-9l7mt\" (UID: \"95067a49-0792-4ee9-a0fd-6cec20b39955\") " pod="openstack/heat-db-create-9l7mt" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.184692 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqvvr\" (UniqueName: \"kubernetes.io/projected/dbca7358-993f-4558-b9b3-e8f41abfbcdf-kube-api-access-xqvvr\") pod \"heat-eaba-account-create-wdpst\" (UID: \"dbca7358-993f-4558-b9b3-e8f41abfbcdf\") " pod="openstack/heat-eaba-account-create-wdpst" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.184740 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd9ss\" (UniqueName: \"kubernetes.io/projected/95067a49-0792-4ee9-a0fd-6cec20b39955-kube-api-access-xd9ss\") pod \"heat-db-create-9l7mt\" (UID: \"95067a49-0792-4ee9-a0fd-6cec20b39955\") " pod="openstack/heat-db-create-9l7mt" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.185947 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95067a49-0792-4ee9-a0fd-6cec20b39955-operator-scripts\") pod \"heat-db-create-9l7mt\" (UID: \"95067a49-0792-4ee9-a0fd-6cec20b39955\") " pod="openstack/heat-db-create-9l7mt" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.201637 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd9ss\" (UniqueName: \"kubernetes.io/projected/95067a49-0792-4ee9-a0fd-6cec20b39955-kube-api-access-xd9ss\") pod \"heat-db-create-9l7mt\" (UID: \"95067a49-0792-4ee9-a0fd-6cec20b39955\") " pod="openstack/heat-db-create-9l7mt" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.286272 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbca7358-993f-4558-b9b3-e8f41abfbcdf-operator-scripts\") pod \"heat-eaba-account-create-wdpst\" (UID: \"dbca7358-993f-4558-b9b3-e8f41abfbcdf\") " pod="openstack/heat-eaba-account-create-wdpst" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.286639 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqvvr\" (UniqueName: \"kubernetes.io/projected/dbca7358-993f-4558-b9b3-e8f41abfbcdf-kube-api-access-xqvvr\") pod \"heat-eaba-account-create-wdpst\" (UID: \"dbca7358-993f-4558-b9b3-e8f41abfbcdf\") " pod="openstack/heat-eaba-account-create-wdpst" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.286886 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbca7358-993f-4558-b9b3-e8f41abfbcdf-operator-scripts\") pod \"heat-eaba-account-create-wdpst\" (UID: \"dbca7358-993f-4558-b9b3-e8f41abfbcdf\") " pod="openstack/heat-eaba-account-create-wdpst" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.303263 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqvvr\" (UniqueName: \"kubernetes.io/projected/dbca7358-993f-4558-b9b3-e8f41abfbcdf-kube-api-access-xqvvr\") pod \"heat-eaba-account-create-wdpst\" (UID: \"dbca7358-993f-4558-b9b3-e8f41abfbcdf\") " pod="openstack/heat-eaba-account-create-wdpst" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.339421 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-9l7mt" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.432459 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-eaba-account-create-wdpst" Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.749533 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74876d679c-sbxz5" event={"ID":"946e72b9-d3e9-4823-8b87-2049e4a107ac","Type":"ContainerStarted","Data":"e106cdf781ed1fb6c12b73aaa0696165940dc433932e5ec6910c8fe16d3c2a47"} Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.749837 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74876d679c-sbxz5" event={"ID":"946e72b9-d3e9-4823-8b87-2049e4a107ac","Type":"ContainerStarted","Data":"8fe22ae8e3fc010ae3ae8a6a44b187eccce819d36740e69cf50891603d261c04"} Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.775423 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-74876d679c-sbxz5" podStartSLOduration=2.775399178 podStartE2EDuration="2.775399178s" podCreationTimestamp="2025-11-24 10:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:29:40.769492929 +0000 UTC m=+5841.303933391" watchObservedRunningTime="2025-11-24 10:29:40.775399178 +0000 UTC m=+5841.309839640" Nov 24 10:29:40 crc kubenswrapper[4944]: W1124 10:29:40.793319 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95067a49_0792_4ee9_a0fd_6cec20b39955.slice/crio-f4f9cebe99b463328a2380fca9841b36731719ea1bc7f00f6c396fe27fd37c7a WatchSource:0}: Error finding container f4f9cebe99b463328a2380fca9841b36731719ea1bc7f00f6c396fe27fd37c7a: Status 404 returned error can't find the container with id f4f9cebe99b463328a2380fca9841b36731719ea1bc7f00f6c396fe27fd37c7a Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.799838 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-9l7mt"] Nov 24 10:29:40 crc kubenswrapper[4944]: I1124 10:29:40.929075 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-eaba-account-create-wdpst"] Nov 24 10:29:40 crc kubenswrapper[4944]: W1124 10:29:40.929583 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbca7358_993f_4558_b9b3_e8f41abfbcdf.slice/crio-b951d190f80626179a210f08c3c2840f9e20d6a896d72c9c582f036bf61f0748 WatchSource:0}: Error finding container b951d190f80626179a210f08c3c2840f9e20d6a896d72c9c582f036bf61f0748: Status 404 returned error can't find the container with id b951d190f80626179a210f08c3c2840f9e20d6a896d72c9c582f036bf61f0748 Nov 24 10:29:41 crc kubenswrapper[4944]: I1124 10:29:41.758776 4944 generic.go:334] "Generic (PLEG): container finished" podID="dbca7358-993f-4558-b9b3-e8f41abfbcdf" containerID="10a3d01cc66611bf337fd55105a30d5361b82dbad1d1968259ab6a4c479f5c45" exitCode=0 Nov 24 10:29:41 crc kubenswrapper[4944]: I1124 10:29:41.759718 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-eaba-account-create-wdpst" event={"ID":"dbca7358-993f-4558-b9b3-e8f41abfbcdf","Type":"ContainerDied","Data":"10a3d01cc66611bf337fd55105a30d5361b82dbad1d1968259ab6a4c479f5c45"} Nov 24 10:29:41 crc kubenswrapper[4944]: I1124 10:29:41.759743 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-eaba-account-create-wdpst" event={"ID":"dbca7358-993f-4558-b9b3-e8f41abfbcdf","Type":"ContainerStarted","Data":"b951d190f80626179a210f08c3c2840f9e20d6a896d72c9c582f036bf61f0748"} Nov 24 10:29:41 crc kubenswrapper[4944]: I1124 10:29:41.761177 4944 generic.go:334] "Generic (PLEG): container finished" podID="95067a49-0792-4ee9-a0fd-6cec20b39955" containerID="53f6c31fbb3ae4d8e49ccc5f3e84a584be9399669eaf919e56a751eb125e09ac" exitCode=0 Nov 24 10:29:41 crc kubenswrapper[4944]: I1124 10:29:41.762291 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-9l7mt" event={"ID":"95067a49-0792-4ee9-a0fd-6cec20b39955","Type":"ContainerDied","Data":"53f6c31fbb3ae4d8e49ccc5f3e84a584be9399669eaf919e56a751eb125e09ac"} Nov 24 10:29:41 crc kubenswrapper[4944]: I1124 10:29:41.762318 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-9l7mt" event={"ID":"95067a49-0792-4ee9-a0fd-6cec20b39955","Type":"ContainerStarted","Data":"f4f9cebe99b463328a2380fca9841b36731719ea1bc7f00f6c396fe27fd37c7a"} Nov 24 10:29:42 crc kubenswrapper[4944]: I1124 10:29:42.039273 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-071a-account-create-b86qt"] Nov 24 10:29:42 crc kubenswrapper[4944]: I1124 10:29:42.049719 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-kwp4n"] Nov 24 10:29:42 crc kubenswrapper[4944]: I1124 10:29:42.061559 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-071a-account-create-b86qt"] Nov 24 10:29:42 crc kubenswrapper[4944]: I1124 10:29:42.070697 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-kwp4n"] Nov 24 10:29:42 crc kubenswrapper[4944]: I1124 10:29:42.286859 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="123e08ab-6f70-4f18-a56b-149b8879fd0c" path="/var/lib/kubelet/pods/123e08ab-6f70-4f18-a56b-149b8879fd0c/volumes" Nov 24 10:29:42 crc kubenswrapper[4944]: I1124 10:29:42.288151 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2da0f3e-f399-432d-81c3-6a20b7483511" path="/var/lib/kubelet/pods/f2da0f3e-f399-432d-81c3-6a20b7483511/volumes" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.198898 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-9l7mt" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.214988 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-eaba-account-create-wdpst" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.360611 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95067a49-0792-4ee9-a0fd-6cec20b39955-operator-scripts\") pod \"95067a49-0792-4ee9-a0fd-6cec20b39955\" (UID: \"95067a49-0792-4ee9-a0fd-6cec20b39955\") " Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.360734 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbca7358-993f-4558-b9b3-e8f41abfbcdf-operator-scripts\") pod \"dbca7358-993f-4558-b9b3-e8f41abfbcdf\" (UID: \"dbca7358-993f-4558-b9b3-e8f41abfbcdf\") " Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.360775 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd9ss\" (UniqueName: \"kubernetes.io/projected/95067a49-0792-4ee9-a0fd-6cec20b39955-kube-api-access-xd9ss\") pod \"95067a49-0792-4ee9-a0fd-6cec20b39955\" (UID: \"95067a49-0792-4ee9-a0fd-6cec20b39955\") " Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.360978 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqvvr\" (UniqueName: \"kubernetes.io/projected/dbca7358-993f-4558-b9b3-e8f41abfbcdf-kube-api-access-xqvvr\") pod \"dbca7358-993f-4558-b9b3-e8f41abfbcdf\" (UID: \"dbca7358-993f-4558-b9b3-e8f41abfbcdf\") " Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.364392 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95067a49-0792-4ee9-a0fd-6cec20b39955-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95067a49-0792-4ee9-a0fd-6cec20b39955" (UID: "95067a49-0792-4ee9-a0fd-6cec20b39955"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.364717 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbca7358-993f-4558-b9b3-e8f41abfbcdf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dbca7358-993f-4558-b9b3-e8f41abfbcdf" (UID: "dbca7358-993f-4558-b9b3-e8f41abfbcdf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.370293 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbca7358-993f-4558-b9b3-e8f41abfbcdf-kube-api-access-xqvvr" (OuterVolumeSpecName: "kube-api-access-xqvvr") pod "dbca7358-993f-4558-b9b3-e8f41abfbcdf" (UID: "dbca7358-993f-4558-b9b3-e8f41abfbcdf"). InnerVolumeSpecName "kube-api-access-xqvvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.370338 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95067a49-0792-4ee9-a0fd-6cec20b39955-kube-api-access-xd9ss" (OuterVolumeSpecName: "kube-api-access-xd9ss") pod "95067a49-0792-4ee9-a0fd-6cec20b39955" (UID: "95067a49-0792-4ee9-a0fd-6cec20b39955"). InnerVolumeSpecName "kube-api-access-xd9ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.463058 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95067a49-0792-4ee9-a0fd-6cec20b39955-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.463084 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dbca7358-993f-4558-b9b3-e8f41abfbcdf-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.463094 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd9ss\" (UniqueName: \"kubernetes.io/projected/95067a49-0792-4ee9-a0fd-6cec20b39955-kube-api-access-xd9ss\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.463104 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqvvr\" (UniqueName: \"kubernetes.io/projected/dbca7358-993f-4558-b9b3-e8f41abfbcdf-kube-api-access-xqvvr\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.787996 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-9l7mt" event={"ID":"95067a49-0792-4ee9-a0fd-6cec20b39955","Type":"ContainerDied","Data":"f4f9cebe99b463328a2380fca9841b36731719ea1bc7f00f6c396fe27fd37c7a"} Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.788031 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-9l7mt" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.788038 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4f9cebe99b463328a2380fca9841b36731719ea1bc7f00f6c396fe27fd37c7a" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.799435 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-eaba-account-create-wdpst" event={"ID":"dbca7358-993f-4558-b9b3-e8f41abfbcdf","Type":"ContainerDied","Data":"b951d190f80626179a210f08c3c2840f9e20d6a896d72c9c582f036bf61f0748"} Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.799471 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-eaba-account-create-wdpst" Nov 24 10:29:43 crc kubenswrapper[4944]: I1124 10:29:43.799479 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b951d190f80626179a210f08c3c2840f9e20d6a896d72c9c582f036bf61f0748" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.218717 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-szw7j"] Nov 24 10:29:45 crc kubenswrapper[4944]: E1124 10:29:45.219617 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbca7358-993f-4558-b9b3-e8f41abfbcdf" containerName="mariadb-account-create" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.219634 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbca7358-993f-4558-b9b3-e8f41abfbcdf" containerName="mariadb-account-create" Nov 24 10:29:45 crc kubenswrapper[4944]: E1124 10:29:45.219668 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95067a49-0792-4ee9-a0fd-6cec20b39955" containerName="mariadb-database-create" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.219674 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="95067a49-0792-4ee9-a0fd-6cec20b39955" containerName="mariadb-database-create" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.219858 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbca7358-993f-4558-b9b3-e8f41abfbcdf" containerName="mariadb-account-create" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.219882 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="95067a49-0792-4ee9-a0fd-6cec20b39955" containerName="mariadb-database-create" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.220625 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.222421 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-5f7rv" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.223571 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.234682 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-szw7j"] Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.297376 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-combined-ca-bundle\") pod \"heat-db-sync-szw7j\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.297539 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-config-data\") pod \"heat-db-sync-szw7j\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.297567 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twf9q\" (UniqueName: \"kubernetes.io/projected/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-kube-api-access-twf9q\") pod \"heat-db-sync-szw7j\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.399230 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-config-data\") pod \"heat-db-sync-szw7j\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.399282 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twf9q\" (UniqueName: \"kubernetes.io/projected/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-kube-api-access-twf9q\") pod \"heat-db-sync-szw7j\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.399379 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-combined-ca-bundle\") pod \"heat-db-sync-szw7j\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.404794 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-config-data\") pod \"heat-db-sync-szw7j\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.415917 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-combined-ca-bundle\") pod \"heat-db-sync-szw7j\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.416034 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twf9q\" (UniqueName: \"kubernetes.io/projected/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-kube-api-access-twf9q\") pod \"heat-db-sync-szw7j\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:45 crc kubenswrapper[4944]: I1124 10:29:45.547847 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:46 crc kubenswrapper[4944]: I1124 10:29:46.068324 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-szw7j"] Nov 24 10:29:46 crc kubenswrapper[4944]: I1124 10:29:46.070545 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 10:29:46 crc kubenswrapper[4944]: I1124 10:29:46.827645 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-szw7j" event={"ID":"8bc364f9-82a1-4f2f-9eed-980aa56e8e57","Type":"ContainerStarted","Data":"e8b06c873819712ba81ee78b293f0fbc20af8fcaa2d0858e8c75f5373c29d21c"} Nov 24 10:29:47 crc kubenswrapper[4944]: I1124 10:29:47.277242 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:29:47 crc kubenswrapper[4944]: E1124 10:29:47.277466 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:29:47 crc kubenswrapper[4944]: I1124 10:29:47.448883 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-69b6d48bbf-jwqrd" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Nov 24 10:29:49 crc kubenswrapper[4944]: I1124 10:29:49.161507 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:49 crc kubenswrapper[4944]: I1124 10:29:49.162079 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:29:51 crc kubenswrapper[4944]: I1124 10:29:51.038474 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-mnjjq"] Nov 24 10:29:51 crc kubenswrapper[4944]: I1124 10:29:51.071213 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-mnjjq"] Nov 24 10:29:52 crc kubenswrapper[4944]: I1124 10:29:52.288027 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06c01ef8-c21d-47a9-80ef-58ade249d002" path="/var/lib/kubelet/pods/06c01ef8-c21d-47a9-80ef-58ade249d002/volumes" Nov 24 10:29:52 crc kubenswrapper[4944]: I1124 10:29:52.896962 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-szw7j" event={"ID":"8bc364f9-82a1-4f2f-9eed-980aa56e8e57","Type":"ContainerStarted","Data":"f4f7775af7d56e1d996b230ef9d8a67199ec7f3c69c55cc3f8ca797cc1fc289e"} Nov 24 10:29:52 crc kubenswrapper[4944]: I1124 10:29:52.916013 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-szw7j" podStartSLOduration=1.652197561 podStartE2EDuration="7.91599324s" podCreationTimestamp="2025-11-24 10:29:45 +0000 UTC" firstStartedPulling="2025-11-24 10:29:46.070342619 +0000 UTC m=+5846.604783081" lastFinishedPulling="2025-11-24 10:29:52.334138298 +0000 UTC m=+5852.868578760" observedRunningTime="2025-11-24 10:29:52.910308099 +0000 UTC m=+5853.444748561" watchObservedRunningTime="2025-11-24 10:29:52.91599324 +0000 UTC m=+5853.450433702" Nov 24 10:29:54 crc kubenswrapper[4944]: I1124 10:29:54.914981 4944 generic.go:334] "Generic (PLEG): container finished" podID="8bc364f9-82a1-4f2f-9eed-980aa56e8e57" containerID="f4f7775af7d56e1d996b230ef9d8a67199ec7f3c69c55cc3f8ca797cc1fc289e" exitCode=0 Nov 24 10:29:54 crc kubenswrapper[4944]: I1124 10:29:54.915109 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-szw7j" event={"ID":"8bc364f9-82a1-4f2f-9eed-980aa56e8e57","Type":"ContainerDied","Data":"f4f7775af7d56e1d996b230ef9d8a67199ec7f3c69c55cc3f8ca797cc1fc289e"} Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.269129 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.355536 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-combined-ca-bundle\") pod \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.355849 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-config-data\") pod \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.355897 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twf9q\" (UniqueName: \"kubernetes.io/projected/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-kube-api-access-twf9q\") pod \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\" (UID: \"8bc364f9-82a1-4f2f-9eed-980aa56e8e57\") " Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.363222 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-kube-api-access-twf9q" (OuterVolumeSpecName: "kube-api-access-twf9q") pod "8bc364f9-82a1-4f2f-9eed-980aa56e8e57" (UID: "8bc364f9-82a1-4f2f-9eed-980aa56e8e57"). InnerVolumeSpecName "kube-api-access-twf9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.393396 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bc364f9-82a1-4f2f-9eed-980aa56e8e57" (UID: "8bc364f9-82a1-4f2f-9eed-980aa56e8e57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.448581 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-config-data" (OuterVolumeSpecName: "config-data") pod "8bc364f9-82a1-4f2f-9eed-980aa56e8e57" (UID: "8bc364f9-82a1-4f2f-9eed-980aa56e8e57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.457743 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twf9q\" (UniqueName: \"kubernetes.io/projected/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-kube-api-access-twf9q\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.457779 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.457788 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bc364f9-82a1-4f2f-9eed-980aa56e8e57-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.940031 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-szw7j" event={"ID":"8bc364f9-82a1-4f2f-9eed-980aa56e8e57","Type":"ContainerDied","Data":"e8b06c873819712ba81ee78b293f0fbc20af8fcaa2d0858e8c75f5373c29d21c"} Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.940091 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8b06c873819712ba81ee78b293f0fbc20af8fcaa2d0858e8c75f5373c29d21c" Nov 24 10:29:56 crc kubenswrapper[4944]: I1124 10:29:56.940151 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-szw7j" Nov 24 10:29:57 crc kubenswrapper[4944]: I1124 10:29:57.448690 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-69b6d48bbf-jwqrd" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.107:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.107:8080: connect: connection refused" Nov 24 10:29:57 crc kubenswrapper[4944]: I1124 10:29:57.449244 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:29:57 crc kubenswrapper[4944]: I1124 10:29:57.906759 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-9544b586c-rkdrm"] Nov 24 10:29:57 crc kubenswrapper[4944]: E1124 10:29:57.907302 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bc364f9-82a1-4f2f-9eed-980aa56e8e57" containerName="heat-db-sync" Nov 24 10:29:57 crc kubenswrapper[4944]: I1124 10:29:57.907322 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bc364f9-82a1-4f2f-9eed-980aa56e8e57" containerName="heat-db-sync" Nov 24 10:29:57 crc kubenswrapper[4944]: I1124 10:29:57.907570 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bc364f9-82a1-4f2f-9eed-980aa56e8e57" containerName="heat-db-sync" Nov 24 10:29:57 crc kubenswrapper[4944]: I1124 10:29:57.912035 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:57 crc kubenswrapper[4944]: I1124 10:29:57.918145 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Nov 24 10:29:57 crc kubenswrapper[4944]: I1124 10:29:57.918312 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-5f7rv" Nov 24 10:29:57 crc kubenswrapper[4944]: I1124 10:29:57.918470 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 24 10:29:57 crc kubenswrapper[4944]: I1124 10:29:57.930008 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-9544b586c-rkdrm"] Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.058546 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5987fb6488-ftkqz"] Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.060072 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.072726 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.076080 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5987fb6488-ftkqz"] Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.090135 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6d6c5dbc76-jsn6f"] Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.094173 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.096452 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt8cb\" (UniqueName: \"kubernetes.io/projected/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-kube-api-access-lt8cb\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.096521 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430c37b1-3658-4472-aa78-ac54f934d9fb-config-data\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.096631 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-config-data-custom\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.096674 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b47f2\" (UniqueName: \"kubernetes.io/projected/430c37b1-3658-4472-aa78-ac54f934d9fb-kube-api-access-b47f2\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.096729 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-combined-ca-bundle\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.096777 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/430c37b1-3658-4472-aa78-ac54f934d9fb-config-data-custom\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.096838 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430c37b1-3658-4472-aa78-ac54f934d9fb-combined-ca-bundle\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.096885 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-config-data\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.097536 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.140813 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6d6c5dbc76-jsn6f"] Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198561 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430c37b1-3658-4472-aa78-ac54f934d9fb-combined-ca-bundle\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198616 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-config-data\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198645 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt8cb\" (UniqueName: \"kubernetes.io/projected/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-kube-api-access-lt8cb\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198669 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f750eb4-b853-445a-acd3-dece22f6bcf3-combined-ca-bundle\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198694 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430c37b1-3658-4472-aa78-ac54f934d9fb-config-data\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198730 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f750eb4-b853-445a-acd3-dece22f6bcf3-config-data\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198771 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-config-data-custom\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198823 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b47f2\" (UniqueName: \"kubernetes.io/projected/430c37b1-3658-4472-aa78-ac54f934d9fb-kube-api-access-b47f2\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198860 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-combined-ca-bundle\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198877 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f750eb4-b853-445a-acd3-dece22f6bcf3-config-data-custom\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198906 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/430c37b1-3658-4472-aa78-ac54f934d9fb-config-data-custom\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.198939 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgfvs\" (UniqueName: \"kubernetes.io/projected/7f750eb4-b853-445a-acd3-dece22f6bcf3-kube-api-access-jgfvs\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.228345 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt8cb\" (UniqueName: \"kubernetes.io/projected/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-kube-api-access-lt8cb\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.232185 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430c37b1-3658-4472-aa78-ac54f934d9fb-combined-ca-bundle\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.232421 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-config-data\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.233153 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-config-data-custom\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.240363 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430c37b1-3658-4472-aa78-ac54f934d9fb-config-data\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.248928 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/430c37b1-3658-4472-aa78-ac54f934d9fb-config-data-custom\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.249788 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9-combined-ca-bundle\") pod \"heat-cfnapi-6d6c5dbc76-jsn6f\" (UID: \"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9\") " pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.248887 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b47f2\" (UniqueName: \"kubernetes.io/projected/430c37b1-3658-4472-aa78-ac54f934d9fb-kube-api-access-b47f2\") pod \"heat-engine-9544b586c-rkdrm\" (UID: \"430c37b1-3658-4472-aa78-ac54f934d9fb\") " pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.300780 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f750eb4-b853-445a-acd3-dece22f6bcf3-config-data-custom\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.317673 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgfvs\" (UniqueName: \"kubernetes.io/projected/7f750eb4-b853-445a-acd3-dece22f6bcf3-kube-api-access-jgfvs\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.318023 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f750eb4-b853-445a-acd3-dece22f6bcf3-combined-ca-bundle\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.318300 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f750eb4-b853-445a-acd3-dece22f6bcf3-config-data\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.308979 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f750eb4-b853-445a-acd3-dece22f6bcf3-config-data-custom\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.323160 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f750eb4-b853-445a-acd3-dece22f6bcf3-combined-ca-bundle\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.338572 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f750eb4-b853-445a-acd3-dece22f6bcf3-config-data\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.377858 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgfvs\" (UniqueName: \"kubernetes.io/projected/7f750eb4-b853-445a-acd3-dece22f6bcf3-kube-api-access-jgfvs\") pod \"heat-api-5987fb6488-ftkqz\" (UID: \"7f750eb4-b853-445a-acd3-dece22f6bcf3\") " pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.391266 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.428495 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:29:58 crc kubenswrapper[4944]: I1124 10:29:58.552136 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:29:59 crc kubenswrapper[4944]: I1124 10:29:59.164380 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-74876d679c-sbxz5" podUID="946e72b9-d3e9-4823-8b87-2049e4a107ac" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.111:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.111:8080: connect: connection refused" Nov 24 10:29:59 crc kubenswrapper[4944]: I1124 10:29:59.516444 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-9544b586c-rkdrm"] Nov 24 10:29:59 crc kubenswrapper[4944]: I1124 10:29:59.553113 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6d6c5dbc76-jsn6f"] Nov 24 10:29:59 crc kubenswrapper[4944]: I1124 10:29:59.683710 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5987fb6488-ftkqz"] Nov 24 10:29:59 crc kubenswrapper[4944]: W1124 10:29:59.702606 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f750eb4_b853_445a_acd3_dece22f6bcf3.slice/crio-46c4d083cbac19257b3c69aa3e3b1afb0c07699e5a5e2196b1714c056f92201e WatchSource:0}: Error finding container 46c4d083cbac19257b3c69aa3e3b1afb0c07699e5a5e2196b1714c056f92201e: Status 404 returned error can't find the container with id 46c4d083cbac19257b3c69aa3e3b1afb0c07699e5a5e2196b1714c056f92201e Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.121457 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-9544b586c-rkdrm" event={"ID":"430c37b1-3658-4472-aa78-ac54f934d9fb","Type":"ContainerStarted","Data":"b7a3a8360c8c56775b629bfb4016627eda409ac2e5d42ec75f0b247e88e74ed7"} Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.121596 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-9544b586c-rkdrm" event={"ID":"430c37b1-3658-4472-aa78-ac54f934d9fb","Type":"ContainerStarted","Data":"114d2dda1d4c9588911a683bcbd17cbc34e09efa0c5cf87d4a7a611147c31035"} Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.121633 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.130414 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" event={"ID":"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9","Type":"ContainerStarted","Data":"b1a4cd355d117d857f4d30e7901b7fe5cabad6e4a88e05b56d14dffe8cfd8671"} Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.145880 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5987fb6488-ftkqz" event={"ID":"7f750eb4-b853-445a-acd3-dece22f6bcf3","Type":"ContainerStarted","Data":"46c4d083cbac19257b3c69aa3e3b1afb0c07699e5a5e2196b1714c056f92201e"} Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.160819 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm"] Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.163664 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.167605 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.167899 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.173608 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-9544b586c-rkdrm" podStartSLOduration=3.173572561 podStartE2EDuration="3.173572561s" podCreationTimestamp="2025-11-24 10:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:30:00.152350834 +0000 UTC m=+5860.686791306" watchObservedRunningTime="2025-11-24 10:30:00.173572561 +0000 UTC m=+5860.708013023" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.191951 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm"] Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.272819 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p7gx\" (UniqueName: \"kubernetes.io/projected/ac6a4cc0-e137-401b-92c0-3835c028d5fa-kube-api-access-5p7gx\") pod \"collect-profiles-29399670-4lwmm\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.272927 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac6a4cc0-e137-401b-92c0-3835c028d5fa-secret-volume\") pod \"collect-profiles-29399670-4lwmm\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.272962 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac6a4cc0-e137-401b-92c0-3835c028d5fa-config-volume\") pod \"collect-profiles-29399670-4lwmm\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.291942 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:30:00 crc kubenswrapper[4944]: E1124 10:30:00.292283 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.374857 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac6a4cc0-e137-401b-92c0-3835c028d5fa-config-volume\") pod \"collect-profiles-29399670-4lwmm\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.375093 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p7gx\" (UniqueName: \"kubernetes.io/projected/ac6a4cc0-e137-401b-92c0-3835c028d5fa-kube-api-access-5p7gx\") pod \"collect-profiles-29399670-4lwmm\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.375164 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac6a4cc0-e137-401b-92c0-3835c028d5fa-secret-volume\") pod \"collect-profiles-29399670-4lwmm\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.376123 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac6a4cc0-e137-401b-92c0-3835c028d5fa-config-volume\") pod \"collect-profiles-29399670-4lwmm\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.380137 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac6a4cc0-e137-401b-92c0-3835c028d5fa-secret-volume\") pod \"collect-profiles-29399670-4lwmm\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.404372 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p7gx\" (UniqueName: \"kubernetes.io/projected/ac6a4cc0-e137-401b-92c0-3835c028d5fa-kube-api-access-5p7gx\") pod \"collect-profiles-29399670-4lwmm\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:00 crc kubenswrapper[4944]: I1124 10:30:00.588705 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:01 crc kubenswrapper[4944]: I1124 10:30:01.126313 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm"] Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.250285 4944 generic.go:334] "Generic (PLEG): container finished" podID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerID="5955542405b349aeb8d0ef4b9ddea8cd5b9be5b6644b129f0159d056b4b130ef" exitCode=137 Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.251188 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b6d48bbf-jwqrd" event={"ID":"573d39f0-12fd-404e-8c47-cd39e153cbe2","Type":"ContainerDied","Data":"5955542405b349aeb8d0ef4b9ddea8cd5b9be5b6644b129f0159d056b4b130ef"} Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.369386 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" event={"ID":"ac6a4cc0-e137-401b-92c0-3835c028d5fa","Type":"ContainerStarted","Data":"5cac9eb54741fbec54d741c79dd0bffb04a8e6bd91041655a19a9fc8c11cff9c"} Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.647709 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.724767 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/573d39f0-12fd-404e-8c47-cd39e153cbe2-logs\") pod \"573d39f0-12fd-404e-8c47-cd39e153cbe2\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.724847 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-config-data\") pod \"573d39f0-12fd-404e-8c47-cd39e153cbe2\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.724893 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/573d39f0-12fd-404e-8c47-cd39e153cbe2-horizon-secret-key\") pod \"573d39f0-12fd-404e-8c47-cd39e153cbe2\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.724930 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-scripts\") pod \"573d39f0-12fd-404e-8c47-cd39e153cbe2\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.725125 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66x67\" (UniqueName: \"kubernetes.io/projected/573d39f0-12fd-404e-8c47-cd39e153cbe2-kube-api-access-66x67\") pod \"573d39f0-12fd-404e-8c47-cd39e153cbe2\" (UID: \"573d39f0-12fd-404e-8c47-cd39e153cbe2\") " Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.725344 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/573d39f0-12fd-404e-8c47-cd39e153cbe2-logs" (OuterVolumeSpecName: "logs") pod "573d39f0-12fd-404e-8c47-cd39e153cbe2" (UID: "573d39f0-12fd-404e-8c47-cd39e153cbe2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.734633 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/573d39f0-12fd-404e-8c47-cd39e153cbe2-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.734864 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/573d39f0-12fd-404e-8c47-cd39e153cbe2-kube-api-access-66x67" (OuterVolumeSpecName: "kube-api-access-66x67") pod "573d39f0-12fd-404e-8c47-cd39e153cbe2" (UID: "573d39f0-12fd-404e-8c47-cd39e153cbe2"). InnerVolumeSpecName "kube-api-access-66x67". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.761390 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/573d39f0-12fd-404e-8c47-cd39e153cbe2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "573d39f0-12fd-404e-8c47-cd39e153cbe2" (UID: "573d39f0-12fd-404e-8c47-cd39e153cbe2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.765703 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-scripts" (OuterVolumeSpecName: "scripts") pod "573d39f0-12fd-404e-8c47-cd39e153cbe2" (UID: "573d39f0-12fd-404e-8c47-cd39e153cbe2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.783566 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-config-data" (OuterVolumeSpecName: "config-data") pod "573d39f0-12fd-404e-8c47-cd39e153cbe2" (UID: "573d39f0-12fd-404e-8c47-cd39e153cbe2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.837088 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.837130 4944 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/573d39f0-12fd-404e-8c47-cd39e153cbe2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.837161 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/573d39f0-12fd-404e-8c47-cd39e153cbe2-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:02 crc kubenswrapper[4944]: I1124 10:30:02.837174 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66x67\" (UniqueName: \"kubernetes.io/projected/573d39f0-12fd-404e-8c47-cd39e153cbe2-kube-api-access-66x67\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.300647 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69b6d48bbf-jwqrd" event={"ID":"573d39f0-12fd-404e-8c47-cd39e153cbe2","Type":"ContainerDied","Data":"e895723c9019b2dcd23c2dc12caea30b43bc01f8fe5bf63b865c04fc4aa4348f"} Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.300718 4944 scope.go:117] "RemoveContainer" containerID="88243b62391b15a2c3a44a986dfe608b5457e5c130bbf60882723f986e18d386" Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.300839 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69b6d48bbf-jwqrd" Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.307154 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5987fb6488-ftkqz" event={"ID":"7f750eb4-b853-445a-acd3-dece22f6bcf3","Type":"ContainerStarted","Data":"b57266ea30dc218dd2a5db55c1a127735e9a6c582b50296f929a662622fb2b5d"} Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.308142 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.310361 4944 generic.go:334] "Generic (PLEG): container finished" podID="ac6a4cc0-e137-401b-92c0-3835c028d5fa" containerID="b4c0e8a0e7791002aa41ea42b6a8d15bbfac598725e2b5cb07cbeaacf90749f5" exitCode=0 Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.310429 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" event={"ID":"ac6a4cc0-e137-401b-92c0-3835c028d5fa","Type":"ContainerDied","Data":"b4c0e8a0e7791002aa41ea42b6a8d15bbfac598725e2b5cb07cbeaacf90749f5"} Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.319574 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" event={"ID":"d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9","Type":"ContainerStarted","Data":"d5862ab2ff7889887dce0e52d28ffe0bdb0a6bd1d716beb46bf44788f65a0c09"} Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.319765 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.336281 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5987fb6488-ftkqz" podStartSLOduration=2.971782104 podStartE2EDuration="5.336255841s" podCreationTimestamp="2025-11-24 10:29:58 +0000 UTC" firstStartedPulling="2025-11-24 10:29:59.706869397 +0000 UTC m=+5860.241309859" lastFinishedPulling="2025-11-24 10:30:02.071343134 +0000 UTC m=+5862.605783596" observedRunningTime="2025-11-24 10:30:03.32964142 +0000 UTC m=+5863.864081882" watchObservedRunningTime="2025-11-24 10:30:03.336255841 +0000 UTC m=+5863.870696303" Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.373254 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" podStartSLOduration=2.784439686 podStartE2EDuration="5.373229458s" podCreationTimestamp="2025-11-24 10:29:58 +0000 UTC" firstStartedPulling="2025-11-24 10:29:59.522930848 +0000 UTC m=+5860.057371310" lastFinishedPulling="2025-11-24 10:30:02.11172062 +0000 UTC m=+5862.646161082" observedRunningTime="2025-11-24 10:30:03.372150484 +0000 UTC m=+5863.906590946" watchObservedRunningTime="2025-11-24 10:30:03.373229458 +0000 UTC m=+5863.907669920" Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.407753 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69b6d48bbf-jwqrd"] Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.423736 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-69b6d48bbf-jwqrd"] Nov 24 10:30:03 crc kubenswrapper[4944]: I1124 10:30:03.522896 4944 scope.go:117] "RemoveContainer" containerID="5955542405b349aeb8d0ef4b9ddea8cd5b9be5b6644b129f0159d056b4b130ef" Nov 24 10:30:04 crc kubenswrapper[4944]: I1124 10:30:04.290474 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" path="/var/lib/kubelet/pods/573d39f0-12fd-404e-8c47-cd39e153cbe2/volumes" Nov 24 10:30:04 crc kubenswrapper[4944]: I1124 10:30:04.663707 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:04 crc kubenswrapper[4944]: I1124 10:30:04.818005 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac6a4cc0-e137-401b-92c0-3835c028d5fa-config-volume\") pod \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " Nov 24 10:30:04 crc kubenswrapper[4944]: I1124 10:30:04.818417 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac6a4cc0-e137-401b-92c0-3835c028d5fa-secret-volume\") pod \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " Nov 24 10:30:04 crc kubenswrapper[4944]: I1124 10:30:04.818475 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p7gx\" (UniqueName: \"kubernetes.io/projected/ac6a4cc0-e137-401b-92c0-3835c028d5fa-kube-api-access-5p7gx\") pod \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\" (UID: \"ac6a4cc0-e137-401b-92c0-3835c028d5fa\") " Nov 24 10:30:04 crc kubenswrapper[4944]: I1124 10:30:04.818605 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac6a4cc0-e137-401b-92c0-3835c028d5fa-config-volume" (OuterVolumeSpecName: "config-volume") pod "ac6a4cc0-e137-401b-92c0-3835c028d5fa" (UID: "ac6a4cc0-e137-401b-92c0-3835c028d5fa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:30:04 crc kubenswrapper[4944]: I1124 10:30:04.818867 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac6a4cc0-e137-401b-92c0-3835c028d5fa-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:04 crc kubenswrapper[4944]: I1124 10:30:04.825413 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac6a4cc0-e137-401b-92c0-3835c028d5fa-kube-api-access-5p7gx" (OuterVolumeSpecName: "kube-api-access-5p7gx") pod "ac6a4cc0-e137-401b-92c0-3835c028d5fa" (UID: "ac6a4cc0-e137-401b-92c0-3835c028d5fa"). InnerVolumeSpecName "kube-api-access-5p7gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:30:04 crc kubenswrapper[4944]: I1124 10:30:04.829347 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac6a4cc0-e137-401b-92c0-3835c028d5fa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ac6a4cc0-e137-401b-92c0-3835c028d5fa" (UID: "ac6a4cc0-e137-401b-92c0-3835c028d5fa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:30:04 crc kubenswrapper[4944]: I1124 10:30:04.921748 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ac6a4cc0-e137-401b-92c0-3835c028d5fa-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:04 crc kubenswrapper[4944]: I1124 10:30:04.921812 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p7gx\" (UniqueName: \"kubernetes.io/projected/ac6a4cc0-e137-401b-92c0-3835c028d5fa-kube-api-access-5p7gx\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:05 crc kubenswrapper[4944]: I1124 10:30:05.341466 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" event={"ID":"ac6a4cc0-e137-401b-92c0-3835c028d5fa","Type":"ContainerDied","Data":"5cac9eb54741fbec54d741c79dd0bffb04a8e6bd91041655a19a9fc8c11cff9c"} Nov 24 10:30:05 crc kubenswrapper[4944]: I1124 10:30:05.341553 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cac9eb54741fbec54d741c79dd0bffb04a8e6bd91041655a19a9fc8c11cff9c" Nov 24 10:30:05 crc kubenswrapper[4944]: I1124 10:30:05.342202 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm" Nov 24 10:30:05 crc kubenswrapper[4944]: I1124 10:30:05.725875 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c"] Nov 24 10:30:05 crc kubenswrapper[4944]: I1124 10:30:05.734929 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399625-9tg4c"] Nov 24 10:30:06 crc kubenswrapper[4944]: I1124 10:30:06.289874 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3540edcd-cfcd-41eb-b351-9e749e87ce0b" path="/var/lib/kubelet/pods/3540edcd-cfcd-41eb-b351-9e749e87ce0b/volumes" Nov 24 10:30:09 crc kubenswrapper[4944]: I1124 10:30:09.742413 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-5987fb6488-ftkqz" Nov 24 10:30:09 crc kubenswrapper[4944]: I1124 10:30:09.791280 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-6d6c5dbc76-jsn6f" Nov 24 10:30:11 crc kubenswrapper[4944]: I1124 10:30:11.329161 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:30:13 crc kubenswrapper[4944]: I1124 10:30:13.258000 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-74876d679c-sbxz5" Nov 24 10:30:13 crc kubenswrapper[4944]: I1124 10:30:13.337800 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-549c569ffc-sf7qz"] Nov 24 10:30:13 crc kubenswrapper[4944]: I1124 10:30:13.338072 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-549c569ffc-sf7qz" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon-log" containerID="cri-o://2edd0863e2905254a5e16dea1e193dd7b8f6cc94e96cdb119d18b6c6d963bb41" gracePeriod=30 Nov 24 10:30:13 crc kubenswrapper[4944]: I1124 10:30:13.338539 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-549c569ffc-sf7qz" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon" containerID="cri-o://1155fa3580f692cb5172297de3ebfc0c787f3a83617bff8eeed96ab2d859150d" gracePeriod=30 Nov 24 10:30:15 crc kubenswrapper[4944]: I1124 10:30:15.276937 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:30:15 crc kubenswrapper[4944]: E1124 10:30:15.277549 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:30:17 crc kubenswrapper[4944]: I1124 10:30:17.473517 4944 generic.go:334] "Generic (PLEG): container finished" podID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerID="1155fa3580f692cb5172297de3ebfc0c787f3a83617bff8eeed96ab2d859150d" exitCode=0 Nov 24 10:30:17 crc kubenswrapper[4944]: I1124 10:30:17.473843 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549c569ffc-sf7qz" event={"ID":"cb28ab61-6a32-4fa9-986b-be0b8e69b56d","Type":"ContainerDied","Data":"1155fa3580f692cb5172297de3ebfc0c787f3a83617bff8eeed96ab2d859150d"} Nov 24 10:30:17 crc kubenswrapper[4944]: I1124 10:30:17.960686 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-549c569ffc-sf7qz" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.108:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.108:8080: connect: connection refused" Nov 24 10:30:18 crc kubenswrapper[4944]: I1124 10:30:18.586970 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-9544b586c-rkdrm" Nov 24 10:30:26 crc kubenswrapper[4944]: I1124 10:30:26.277258 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:30:26 crc kubenswrapper[4944]: E1124 10:30:26.278192 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:30:27 crc kubenswrapper[4944]: I1124 10:30:27.960642 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-549c569ffc-sf7qz" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.108:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.108:8080: connect: connection refused" Nov 24 10:30:31 crc kubenswrapper[4944]: I1124 10:30:31.939317 4944 scope.go:117] "RemoveContainer" containerID="f4b8e477abf074b965c95928aed2ca799c57b9974022dc800dace88373936414" Nov 24 10:30:31 crc kubenswrapper[4944]: I1124 10:30:31.990368 4944 scope.go:117] "RemoveContainer" containerID="ec46b567ec5387c476237532f4c35144e9bf880ec8434b125bcedaa596d38894" Nov 24 10:30:32 crc kubenswrapper[4944]: I1124 10:30:32.031224 4944 scope.go:117] "RemoveContainer" containerID="e7280220640a036da261a00ad8167e97d28ed7af3a867d513c67dec04b359dc3" Nov 24 10:30:32 crc kubenswrapper[4944]: I1124 10:30:32.059103 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-2a01-account-create-h4kll"] Nov 24 10:30:32 crc kubenswrapper[4944]: I1124 10:30:32.070239 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-m9kvn"] Nov 24 10:30:32 crc kubenswrapper[4944]: I1124 10:30:32.078931 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-2a01-account-create-h4kll"] Nov 24 10:30:32 crc kubenswrapper[4944]: I1124 10:30:32.085847 4944 scope.go:117] "RemoveContainer" containerID="7b3d325908fb3be09b9dd458e99b0001089c2ef16aa428b2c9210d7857257549" Nov 24 10:30:32 crc kubenswrapper[4944]: I1124 10:30:32.088606 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-m9kvn"] Nov 24 10:30:32 crc kubenswrapper[4944]: I1124 10:30:32.109864 4944 scope.go:117] "RemoveContainer" containerID="6ffb42869327f7cfcd9e56622a185ddbb74cb32fe121b336bd6221ca41b5f711" Nov 24 10:30:32 crc kubenswrapper[4944]: I1124 10:30:32.144118 4944 scope.go:117] "RemoveContainer" containerID="ff0dbb3d0734e85266fb5db92d0779680ea8e630abbc2b0a439bc4195aaa4403" Nov 24 10:30:32 crc kubenswrapper[4944]: I1124 10:30:32.290769 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4abb1c85-a715-4823-ad1e-0a0fbdbae095" path="/var/lib/kubelet/pods/4abb1c85-a715-4823-ad1e-0a0fbdbae095/volumes" Nov 24 10:30:32 crc kubenswrapper[4944]: I1124 10:30:32.292196 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5d435c0-196f-4a8d-89f7-5e4da0de7eed" path="/var/lib/kubelet/pods/b5d435c0-196f-4a8d-89f7-5e4da0de7eed/volumes" Nov 24 10:30:37 crc kubenswrapper[4944]: I1124 10:30:37.961136 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-549c569ffc-sf7qz" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.108:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.108:8080: connect: connection refused" Nov 24 10:30:37 crc kubenswrapper[4944]: I1124 10:30:37.962368 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:30:39 crc kubenswrapper[4944]: I1124 10:30:39.277252 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:30:39 crc kubenswrapper[4944]: E1124 10:30:39.278003 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:30:41 crc kubenswrapper[4944]: I1124 10:30:41.042267 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-r774x"] Nov 24 10:30:41 crc kubenswrapper[4944]: I1124 10:30:41.052323 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-r774x"] Nov 24 10:30:42 crc kubenswrapper[4944]: I1124 10:30:42.292473 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2c072d8-6141-44f7-a3cb-59679ce2d5ba" path="/var/lib/kubelet/pods/b2c072d8-6141-44f7-a3cb-59679ce2d5ba/volumes" Nov 24 10:30:43 crc kubenswrapper[4944]: I1124 10:30:43.761318 4944 generic.go:334] "Generic (PLEG): container finished" podID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerID="2edd0863e2905254a5e16dea1e193dd7b8f6cc94e96cdb119d18b6c6d963bb41" exitCode=137 Nov 24 10:30:43 crc kubenswrapper[4944]: I1124 10:30:43.761652 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549c569ffc-sf7qz" event={"ID":"cb28ab61-6a32-4fa9-986b-be0b8e69b56d","Type":"ContainerDied","Data":"2edd0863e2905254a5e16dea1e193dd7b8f6cc94e96cdb119d18b6c6d963bb41"} Nov 24 10:30:43 crc kubenswrapper[4944]: I1124 10:30:43.975135 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.040805 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-logs\") pod \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.041068 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-scripts\") pod \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.041157 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsbsd\" (UniqueName: \"kubernetes.io/projected/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-kube-api-access-lsbsd\") pod \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.041203 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-horizon-secret-key\") pod \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.041239 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-config-data\") pod \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\" (UID: \"cb28ab61-6a32-4fa9-986b-be0b8e69b56d\") " Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.041309 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-logs" (OuterVolumeSpecName: "logs") pod "cb28ab61-6a32-4fa9-986b-be0b8e69b56d" (UID: "cb28ab61-6a32-4fa9-986b-be0b8e69b56d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.042139 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-logs\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.050311 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-kube-api-access-lsbsd" (OuterVolumeSpecName: "kube-api-access-lsbsd") pod "cb28ab61-6a32-4fa9-986b-be0b8e69b56d" (UID: "cb28ab61-6a32-4fa9-986b-be0b8e69b56d"). InnerVolumeSpecName "kube-api-access-lsbsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.056275 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "cb28ab61-6a32-4fa9-986b-be0b8e69b56d" (UID: "cb28ab61-6a32-4fa9-986b-be0b8e69b56d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.071108 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-config-data" (OuterVolumeSpecName: "config-data") pod "cb28ab61-6a32-4fa9-986b-be0b8e69b56d" (UID: "cb28ab61-6a32-4fa9-986b-be0b8e69b56d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.079655 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-scripts" (OuterVolumeSpecName: "scripts") pod "cb28ab61-6a32-4fa9-986b-be0b8e69b56d" (UID: "cb28ab61-6a32-4fa9-986b-be0b8e69b56d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.145383 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsbsd\" (UniqueName: \"kubernetes.io/projected/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-kube-api-access-lsbsd\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.145429 4944 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.145491 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.145508 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb28ab61-6a32-4fa9-986b-be0b8e69b56d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.777030 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-549c569ffc-sf7qz" event={"ID":"cb28ab61-6a32-4fa9-986b-be0b8e69b56d","Type":"ContainerDied","Data":"8b70bc6924c13cd91aef1cfb433a096481252871374b8baa265df0baa82e26f4"} Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.777539 4944 scope.go:117] "RemoveContainer" containerID="1155fa3580f692cb5172297de3ebfc0c787f3a83617bff8eeed96ab2d859150d" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.777736 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-549c569ffc-sf7qz" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.805734 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-549c569ffc-sf7qz"] Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.813493 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-549c569ffc-sf7qz"] Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.959236 4944 scope.go:117] "RemoveContainer" containerID="2edd0863e2905254a5e16dea1e193dd7b8f6cc94e96cdb119d18b6c6d963bb41" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.981002 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds"] Nov 24 10:30:44 crc kubenswrapper[4944]: E1124 10:30:44.981543 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon-log" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.981563 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon-log" Nov 24 10:30:44 crc kubenswrapper[4944]: E1124 10:30:44.981587 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac6a4cc0-e137-401b-92c0-3835c028d5fa" containerName="collect-profiles" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.981594 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac6a4cc0-e137-401b-92c0-3835c028d5fa" containerName="collect-profiles" Nov 24 10:30:44 crc kubenswrapper[4944]: E1124 10:30:44.981607 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.981615 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon" Nov 24 10:30:44 crc kubenswrapper[4944]: E1124 10:30:44.981646 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon-log" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.981652 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon-log" Nov 24 10:30:44 crc kubenswrapper[4944]: E1124 10:30:44.981661 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.981666 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.981842 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.981856 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac6a4cc0-e137-401b-92c0-3835c028d5fa" containerName="collect-profiles" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.981868 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.981876 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" containerName="horizon-log" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.981892 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="573d39f0-12fd-404e-8c47-cd39e153cbe2" containerName="horizon-log" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.983537 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:44 crc kubenswrapper[4944]: I1124 10:30:44.989853 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.011995 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds"] Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.077460 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.077537 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h6wg\" (UniqueName: \"kubernetes.io/projected/181b11f3-6b16-4cfc-962e-48687b6823c1-kube-api-access-6h6wg\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.077867 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.181079 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.181189 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h6wg\" (UniqueName: \"kubernetes.io/projected/181b11f3-6b16-4cfc-962e-48687b6823c1-kube-api-access-6h6wg\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.181263 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.181632 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.181805 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.208243 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h6wg\" (UniqueName: \"kubernetes.io/projected/181b11f3-6b16-4cfc-962e-48687b6823c1-kube-api-access-6h6wg\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.302136 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:45 crc kubenswrapper[4944]: I1124 10:30:45.806894 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds"] Nov 24 10:30:46 crc kubenswrapper[4944]: I1124 10:30:46.291702 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb28ab61-6a32-4fa9-986b-be0b8e69b56d" path="/var/lib/kubelet/pods/cb28ab61-6a32-4fa9-986b-be0b8e69b56d/volumes" Nov 24 10:30:46 crc kubenswrapper[4944]: I1124 10:30:46.811647 4944 generic.go:334] "Generic (PLEG): container finished" podID="181b11f3-6b16-4cfc-962e-48687b6823c1" containerID="bad3726f7e25c9cdeb58bd6b43af7743269abedb6f5764930803a84b2900fd92" exitCode=0 Nov 24 10:30:46 crc kubenswrapper[4944]: I1124 10:30:46.811740 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" event={"ID":"181b11f3-6b16-4cfc-962e-48687b6823c1","Type":"ContainerDied","Data":"bad3726f7e25c9cdeb58bd6b43af7743269abedb6f5764930803a84b2900fd92"} Nov 24 10:30:46 crc kubenswrapper[4944]: I1124 10:30:46.811785 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" event={"ID":"181b11f3-6b16-4cfc-962e-48687b6823c1","Type":"ContainerStarted","Data":"4591798110d844a2632b9134ce6119c7accc9c7c64e5c6fc38118e1815230778"} Nov 24 10:30:48 crc kubenswrapper[4944]: I1124 10:30:48.853981 4944 generic.go:334] "Generic (PLEG): container finished" podID="181b11f3-6b16-4cfc-962e-48687b6823c1" containerID="da4fee48ed982a7e235dcee0607e2e070d8c8c3d4abf84b9ffa69142f512e6dd" exitCode=0 Nov 24 10:30:48 crc kubenswrapper[4944]: I1124 10:30:48.854078 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" event={"ID":"181b11f3-6b16-4cfc-962e-48687b6823c1","Type":"ContainerDied","Data":"da4fee48ed982a7e235dcee0607e2e070d8c8c3d4abf84b9ffa69142f512e6dd"} Nov 24 10:30:49 crc kubenswrapper[4944]: I1124 10:30:49.874855 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" event={"ID":"181b11f3-6b16-4cfc-962e-48687b6823c1","Type":"ContainerStarted","Data":"4c94f06df8008b51f00a275dad87146b6a5df3ed071ecfa617f8b938d9c47ee2"} Nov 24 10:30:49 crc kubenswrapper[4944]: I1124 10:30:49.905203 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" podStartSLOduration=4.513860607 podStartE2EDuration="5.905008745s" podCreationTimestamp="2025-11-24 10:30:44 +0000 UTC" firstStartedPulling="2025-11-24 10:30:46.815374471 +0000 UTC m=+5907.349814933" lastFinishedPulling="2025-11-24 10:30:48.206522579 +0000 UTC m=+5908.740963071" observedRunningTime="2025-11-24 10:30:49.895624086 +0000 UTC m=+5910.430064548" watchObservedRunningTime="2025-11-24 10:30:49.905008745 +0000 UTC m=+5910.439449217" Nov 24 10:30:50 crc kubenswrapper[4944]: I1124 10:30:50.896970 4944 generic.go:334] "Generic (PLEG): container finished" podID="181b11f3-6b16-4cfc-962e-48687b6823c1" containerID="4c94f06df8008b51f00a275dad87146b6a5df3ed071ecfa617f8b938d9c47ee2" exitCode=0 Nov 24 10:30:50 crc kubenswrapper[4944]: I1124 10:30:50.897023 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" event={"ID":"181b11f3-6b16-4cfc-962e-48687b6823c1","Type":"ContainerDied","Data":"4c94f06df8008b51f00a275dad87146b6a5df3ed071ecfa617f8b938d9c47ee2"} Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.243882 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.396216 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-bundle\") pod \"181b11f3-6b16-4cfc-962e-48687b6823c1\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.396288 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h6wg\" (UniqueName: \"kubernetes.io/projected/181b11f3-6b16-4cfc-962e-48687b6823c1-kube-api-access-6h6wg\") pod \"181b11f3-6b16-4cfc-962e-48687b6823c1\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.396455 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-util\") pod \"181b11f3-6b16-4cfc-962e-48687b6823c1\" (UID: \"181b11f3-6b16-4cfc-962e-48687b6823c1\") " Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.402984 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-bundle" (OuterVolumeSpecName: "bundle") pod "181b11f3-6b16-4cfc-962e-48687b6823c1" (UID: "181b11f3-6b16-4cfc-962e-48687b6823c1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.405517 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/181b11f3-6b16-4cfc-962e-48687b6823c1-kube-api-access-6h6wg" (OuterVolumeSpecName: "kube-api-access-6h6wg") pod "181b11f3-6b16-4cfc-962e-48687b6823c1" (UID: "181b11f3-6b16-4cfc-962e-48687b6823c1"). InnerVolumeSpecName "kube-api-access-6h6wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.413799 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-util" (OuterVolumeSpecName: "util") pod "181b11f3-6b16-4cfc-962e-48687b6823c1" (UID: "181b11f3-6b16-4cfc-962e-48687b6823c1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.499756 4944 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.499815 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h6wg\" (UniqueName: \"kubernetes.io/projected/181b11f3-6b16-4cfc-962e-48687b6823c1-kube-api-access-6h6wg\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.499827 4944 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/181b11f3-6b16-4cfc-962e-48687b6823c1-util\") on node \"crc\" DevicePath \"\"" Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.921344 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" event={"ID":"181b11f3-6b16-4cfc-962e-48687b6823c1","Type":"ContainerDied","Data":"4591798110d844a2632b9134ce6119c7accc9c7c64e5c6fc38118e1815230778"} Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.921399 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4591798110d844a2632b9134ce6119c7accc9c7c64e5c6fc38118e1815230778" Nov 24 10:30:52 crc kubenswrapper[4944]: I1124 10:30:52.921417 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds" Nov 24 10:30:53 crc kubenswrapper[4944]: I1124 10:30:53.278121 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:30:53 crc kubenswrapper[4944]: E1124 10:30:53.278373 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:31:01 crc kubenswrapper[4944]: I1124 10:31:01.939242 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-hsntg"] Nov 24 10:31:01 crc kubenswrapper[4944]: E1124 10:31:01.940366 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181b11f3-6b16-4cfc-962e-48687b6823c1" containerName="util" Nov 24 10:31:01 crc kubenswrapper[4944]: I1124 10:31:01.940385 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="181b11f3-6b16-4cfc-962e-48687b6823c1" containerName="util" Nov 24 10:31:01 crc kubenswrapper[4944]: E1124 10:31:01.940414 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181b11f3-6b16-4cfc-962e-48687b6823c1" containerName="pull" Nov 24 10:31:01 crc kubenswrapper[4944]: I1124 10:31:01.940422 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="181b11f3-6b16-4cfc-962e-48687b6823c1" containerName="pull" Nov 24 10:31:01 crc kubenswrapper[4944]: E1124 10:31:01.940485 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181b11f3-6b16-4cfc-962e-48687b6823c1" containerName="extract" Nov 24 10:31:01 crc kubenswrapper[4944]: I1124 10:31:01.940494 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="181b11f3-6b16-4cfc-962e-48687b6823c1" containerName="extract" Nov 24 10:31:01 crc kubenswrapper[4944]: I1124 10:31:01.940769 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="181b11f3-6b16-4cfc-962e-48687b6823c1" containerName="extract" Nov 24 10:31:01 crc kubenswrapper[4944]: I1124 10:31:01.941873 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-hsntg" Nov 24 10:31:01 crc kubenswrapper[4944]: I1124 10:31:01.945066 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-pvpbs" Nov 24 10:31:01 crc kubenswrapper[4944]: I1124 10:31:01.945146 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 24 10:31:01 crc kubenswrapper[4944]: I1124 10:31:01.945971 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 24 10:31:01 crc kubenswrapper[4944]: I1124 10:31:01.952110 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-hsntg"] Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.037239 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp"] Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.038682 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.041412 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-zs5ft" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.043672 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84"] Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.044638 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.045135 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.071348 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp"] Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.090275 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84"] Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.129114 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pddsn\" (UniqueName: \"kubernetes.io/projected/abaf8d9b-a218-4253-8f20-eacd200299f3-kube-api-access-pddsn\") pod \"obo-prometheus-operator-668cf9dfbb-hsntg\" (UID: \"abaf8d9b-a218-4253-8f20-eacd200299f3\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-hsntg" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.165528 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-vrrsh"] Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.166951 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.169566 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-blc4l" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.174876 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.178133 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-vrrsh"] Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.231532 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8cb28caf-699f-4cb4-8c4b-3967a0c5a275-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp\" (UID: \"8cb28caf-699f-4cb4-8c4b-3967a0c5a275\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.231590 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8cb28caf-699f-4cb4-8c4b-3967a0c5a275-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp\" (UID: \"8cb28caf-699f-4cb4-8c4b-3967a0c5a275\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.231778 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b1c3c4c6-4cf3-4653-8ed0-352ceb700925-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84\" (UID: \"b1c3c4c6-4cf3-4653-8ed0-352ceb700925\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.232017 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pddsn\" (UniqueName: \"kubernetes.io/projected/abaf8d9b-a218-4253-8f20-eacd200299f3-kube-api-access-pddsn\") pod \"obo-prometheus-operator-668cf9dfbb-hsntg\" (UID: \"abaf8d9b-a218-4253-8f20-eacd200299f3\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-hsntg" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.232568 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b1c3c4c6-4cf3-4653-8ed0-352ceb700925-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84\" (UID: \"b1c3c4c6-4cf3-4653-8ed0-352ceb700925\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.257869 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-t7fc7"] Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.259125 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-t7fc7" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.264389 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-q4qfg" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.275025 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-t7fc7"] Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.278146 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pddsn\" (UniqueName: \"kubernetes.io/projected/abaf8d9b-a218-4253-8f20-eacd200299f3-kube-api-access-pddsn\") pod \"obo-prometheus-operator-668cf9dfbb-hsntg\" (UID: \"abaf8d9b-a218-4253-8f20-eacd200299f3\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-hsntg" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.335303 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgzq7\" (UniqueName: \"kubernetes.io/projected/b14f0425-11fb-4ed0-9599-0c08bb518a85-kube-api-access-bgzq7\") pod \"observability-operator-d8bb48f5d-vrrsh\" (UID: \"b14f0425-11fb-4ed0-9599-0c08bb518a85\") " pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.335401 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8cb28caf-699f-4cb4-8c4b-3967a0c5a275-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp\" (UID: \"8cb28caf-699f-4cb4-8c4b-3967a0c5a275\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.335423 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8cb28caf-699f-4cb4-8c4b-3967a0c5a275-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp\" (UID: \"8cb28caf-699f-4cb4-8c4b-3967a0c5a275\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.335464 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b1c3c4c6-4cf3-4653-8ed0-352ceb700925-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84\" (UID: \"b1c3c4c6-4cf3-4653-8ed0-352ceb700925\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.335567 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b1c3c4c6-4cf3-4653-8ed0-352ceb700925-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84\" (UID: \"b1c3c4c6-4cf3-4653-8ed0-352ceb700925\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.335599 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/b14f0425-11fb-4ed0-9599-0c08bb518a85-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-vrrsh\" (UID: \"b14f0425-11fb-4ed0-9599-0c08bb518a85\") " pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.359151 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b1c3c4c6-4cf3-4653-8ed0-352ceb700925-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84\" (UID: \"b1c3c4c6-4cf3-4653-8ed0-352ceb700925\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.359827 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8cb28caf-699f-4cb4-8c4b-3967a0c5a275-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp\" (UID: \"8cb28caf-699f-4cb4-8c4b-3967a0c5a275\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.372576 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b1c3c4c6-4cf3-4653-8ed0-352ceb700925-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84\" (UID: \"b1c3c4c6-4cf3-4653-8ed0-352ceb700925\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.381526 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8cb28caf-699f-4cb4-8c4b-3967a0c5a275-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp\" (UID: \"8cb28caf-699f-4cb4-8c4b-3967a0c5a275\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.388837 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.440450 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/571e96c3-7ec5-4d59-8666-c9427e3d2612-openshift-service-ca\") pod \"perses-operator-5446b9c989-t7fc7\" (UID: \"571e96c3-7ec5-4d59-8666-c9427e3d2612\") " pod="openshift-operators/perses-operator-5446b9c989-t7fc7" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.440494 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggvgx\" (UniqueName: \"kubernetes.io/projected/571e96c3-7ec5-4d59-8666-c9427e3d2612-kube-api-access-ggvgx\") pod \"perses-operator-5446b9c989-t7fc7\" (UID: \"571e96c3-7ec5-4d59-8666-c9427e3d2612\") " pod="openshift-operators/perses-operator-5446b9c989-t7fc7" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.440541 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/b14f0425-11fb-4ed0-9599-0c08bb518a85-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-vrrsh\" (UID: \"b14f0425-11fb-4ed0-9599-0c08bb518a85\") " pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.440581 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgzq7\" (UniqueName: \"kubernetes.io/projected/b14f0425-11fb-4ed0-9599-0c08bb518a85-kube-api-access-bgzq7\") pod \"observability-operator-d8bb48f5d-vrrsh\" (UID: \"b14f0425-11fb-4ed0-9599-0c08bb518a85\") " pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.447226 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/b14f0425-11fb-4ed0-9599-0c08bb518a85-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-vrrsh\" (UID: \"b14f0425-11fb-4ed0-9599-0c08bb518a85\") " pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.482923 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgzq7\" (UniqueName: \"kubernetes.io/projected/b14f0425-11fb-4ed0-9599-0c08bb518a85-kube-api-access-bgzq7\") pod \"observability-operator-d8bb48f5d-vrrsh\" (UID: \"b14f0425-11fb-4ed0-9599-0c08bb518a85\") " pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.487882 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.550590 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/571e96c3-7ec5-4d59-8666-c9427e3d2612-openshift-service-ca\") pod \"perses-operator-5446b9c989-t7fc7\" (UID: \"571e96c3-7ec5-4d59-8666-c9427e3d2612\") " pod="openshift-operators/perses-operator-5446b9c989-t7fc7" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.550690 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggvgx\" (UniqueName: \"kubernetes.io/projected/571e96c3-7ec5-4d59-8666-c9427e3d2612-kube-api-access-ggvgx\") pod \"perses-operator-5446b9c989-t7fc7\" (UID: \"571e96c3-7ec5-4d59-8666-c9427e3d2612\") " pod="openshift-operators/perses-operator-5446b9c989-t7fc7" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.552100 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/571e96c3-7ec5-4d59-8666-c9427e3d2612-openshift-service-ca\") pod \"perses-operator-5446b9c989-t7fc7\" (UID: \"571e96c3-7ec5-4d59-8666-c9427e3d2612\") " pod="openshift-operators/perses-operator-5446b9c989-t7fc7" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.572235 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-hsntg" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.590345 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggvgx\" (UniqueName: \"kubernetes.io/projected/571e96c3-7ec5-4d59-8666-c9427e3d2612-kube-api-access-ggvgx\") pod \"perses-operator-5446b9c989-t7fc7\" (UID: \"571e96c3-7ec5-4d59-8666-c9427e3d2612\") " pod="openshift-operators/perses-operator-5446b9c989-t7fc7" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.636847 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-t7fc7" Nov 24 10:31:02 crc kubenswrapper[4944]: I1124 10:31:02.668845 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp" Nov 24 10:31:03 crc kubenswrapper[4944]: I1124 10:31:03.080828 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84"] Nov 24 10:31:03 crc kubenswrapper[4944]: W1124 10:31:03.083918 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1c3c4c6_4cf3_4653_8ed0_352ceb700925.slice/crio-9f81f806ca92aefe97d25f09486c7d1772d9939d71c7f68dcc12d0f49a713e06 WatchSource:0}: Error finding container 9f81f806ca92aefe97d25f09486c7d1772d9939d71c7f68dcc12d0f49a713e06: Status 404 returned error can't find the container with id 9f81f806ca92aefe97d25f09486c7d1772d9939d71c7f68dcc12d0f49a713e06 Nov 24 10:31:03 crc kubenswrapper[4944]: I1124 10:31:03.289420 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-hsntg"] Nov 24 10:31:03 crc kubenswrapper[4944]: I1124 10:31:03.317135 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-vrrsh"] Nov 24 10:31:03 crc kubenswrapper[4944]: I1124 10:31:03.436220 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-t7fc7"] Nov 24 10:31:03 crc kubenswrapper[4944]: I1124 10:31:03.447130 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp"] Nov 24 10:31:03 crc kubenswrapper[4944]: W1124 10:31:03.450804 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod571e96c3_7ec5_4d59_8666_c9427e3d2612.slice/crio-d24007b729d1cdbaab6dcf60bd6bf6ee6ba5202753f5bf3f5ed65b9ddf822abb WatchSource:0}: Error finding container d24007b729d1cdbaab6dcf60bd6bf6ee6ba5202753f5bf3f5ed65b9ddf822abb: Status 404 returned error can't find the container with id d24007b729d1cdbaab6dcf60bd6bf6ee6ba5202753f5bf3f5ed65b9ddf822abb Nov 24 10:31:04 crc kubenswrapper[4944]: I1124 10:31:04.022300 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp" event={"ID":"8cb28caf-699f-4cb4-8c4b-3967a0c5a275","Type":"ContainerStarted","Data":"6802c471d2560ed557174d8c222a6a2c79d4bf6dbb5e590b39022e9bd28c8942"} Nov 24 10:31:04 crc kubenswrapper[4944]: I1124 10:31:04.023762 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-hsntg" event={"ID":"abaf8d9b-a218-4253-8f20-eacd200299f3","Type":"ContainerStarted","Data":"e30bceea8f11b82c8ad3956228fcc47ffc5208656ff75d7bd1ce21398345ba11"} Nov 24 10:31:04 crc kubenswrapper[4944]: I1124 10:31:04.025262 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84" event={"ID":"b1c3c4c6-4cf3-4653-8ed0-352ceb700925","Type":"ContainerStarted","Data":"9f81f806ca92aefe97d25f09486c7d1772d9939d71c7f68dcc12d0f49a713e06"} Nov 24 10:31:04 crc kubenswrapper[4944]: I1124 10:31:04.026711 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-t7fc7" event={"ID":"571e96c3-7ec5-4d59-8666-c9427e3d2612","Type":"ContainerStarted","Data":"d24007b729d1cdbaab6dcf60bd6bf6ee6ba5202753f5bf3f5ed65b9ddf822abb"} Nov 24 10:31:04 crc kubenswrapper[4944]: I1124 10:31:04.027943 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" event={"ID":"b14f0425-11fb-4ed0-9599-0c08bb518a85","Type":"ContainerStarted","Data":"279bfcdd47e20ed2db6bf037b31e39e8f9ee8effd056e38a5323e3d535fea587"} Nov 24 10:31:05 crc kubenswrapper[4944]: I1124 10:31:05.277732 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:31:05 crc kubenswrapper[4944]: E1124 10:31:05.278289 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:31:09 crc kubenswrapper[4944]: I1124 10:31:09.133024 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp" event={"ID":"8cb28caf-699f-4cb4-8c4b-3967a0c5a275","Type":"ContainerStarted","Data":"6aa275058fca2cde597deb66764e4ee4f4ea84516d4fa6edf33b2a99981e622b"} Nov 24 10:31:09 crc kubenswrapper[4944]: I1124 10:31:09.141073 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-hsntg" event={"ID":"abaf8d9b-a218-4253-8f20-eacd200299f3","Type":"ContainerStarted","Data":"1d16203093271d7f5299b89477623c8e7b8bc40a8e96638559ae22a7d1f0bb61"} Nov 24 10:31:09 crc kubenswrapper[4944]: I1124 10:31:09.142820 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84" event={"ID":"b1c3c4c6-4cf3-4653-8ed0-352ceb700925","Type":"ContainerStarted","Data":"a69bb17773f8f35010ef98d22d8f7402dce81e38b645fc1ea90b02af37ab5079"} Nov 24 10:31:09 crc kubenswrapper[4944]: I1124 10:31:09.148196 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-t7fc7" event={"ID":"571e96c3-7ec5-4d59-8666-c9427e3d2612","Type":"ContainerStarted","Data":"1c28bcb5409f5a5f2d471ec292f64698f5e11b481df05b429c64e8fa497b3ee1"} Nov 24 10:31:09 crc kubenswrapper[4944]: I1124 10:31:09.149322 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-t7fc7" Nov 24 10:31:09 crc kubenswrapper[4944]: I1124 10:31:09.157121 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp" podStartSLOduration=3.684395971 podStartE2EDuration="8.157100044s" podCreationTimestamp="2025-11-24 10:31:01 +0000 UTC" firstStartedPulling="2025-11-24 10:31:03.460698847 +0000 UTC m=+5923.995139309" lastFinishedPulling="2025-11-24 10:31:07.93340292 +0000 UTC m=+5928.467843382" observedRunningTime="2025-11-24 10:31:09.15384718 +0000 UTC m=+5929.688287662" watchObservedRunningTime="2025-11-24 10:31:09.157100044 +0000 UTC m=+5929.691540506" Nov 24 10:31:09 crc kubenswrapper[4944]: I1124 10:31:09.249562 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-hsntg" podStartSLOduration=3.638580002 podStartE2EDuration="8.249532378s" podCreationTimestamp="2025-11-24 10:31:01 +0000 UTC" firstStartedPulling="2025-11-24 10:31:03.332614338 +0000 UTC m=+5923.867054790" lastFinishedPulling="2025-11-24 10:31:07.943566704 +0000 UTC m=+5928.478007166" observedRunningTime="2025-11-24 10:31:09.178012211 +0000 UTC m=+5929.712452683" watchObservedRunningTime="2025-11-24 10:31:09.249532378 +0000 UTC m=+5929.783972840" Nov 24 10:31:09 crc kubenswrapper[4944]: I1124 10:31:09.298972 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-t7fc7" podStartSLOduration=2.796053167 podStartE2EDuration="7.298950452s" podCreationTimestamp="2025-11-24 10:31:02 +0000 UTC" firstStartedPulling="2025-11-24 10:31:03.455958216 +0000 UTC m=+5923.990398678" lastFinishedPulling="2025-11-24 10:31:07.958855501 +0000 UTC m=+5928.493295963" observedRunningTime="2025-11-24 10:31:09.207997345 +0000 UTC m=+5929.742437827" watchObservedRunningTime="2025-11-24 10:31:09.298950452 +0000 UTC m=+5929.833390914" Nov 24 10:31:09 crc kubenswrapper[4944]: I1124 10:31:09.342561 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84" podStartSLOduration=3.488742018 podStartE2EDuration="8.34253857s" podCreationTimestamp="2025-11-24 10:31:01 +0000 UTC" firstStartedPulling="2025-11-24 10:31:03.086442526 +0000 UTC m=+5923.620882988" lastFinishedPulling="2025-11-24 10:31:07.940239078 +0000 UTC m=+5928.474679540" observedRunningTime="2025-11-24 10:31:09.233550009 +0000 UTC m=+5929.767990471" watchObservedRunningTime="2025-11-24 10:31:09.34253857 +0000 UTC m=+5929.876979032" Nov 24 10:31:10 crc kubenswrapper[4944]: I1124 10:31:10.070825 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-27fnl"] Nov 24 10:31:10 crc kubenswrapper[4944]: I1124 10:31:10.084255 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-74df-account-create-dmhxf"] Nov 24 10:31:10 crc kubenswrapper[4944]: I1124 10:31:10.095544 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-27fnl"] Nov 24 10:31:10 crc kubenswrapper[4944]: I1124 10:31:10.109139 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-74df-account-create-dmhxf"] Nov 24 10:31:10 crc kubenswrapper[4944]: I1124 10:31:10.301330 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23161bdc-eb83-4926-a369-8a61d912b308" path="/var/lib/kubelet/pods/23161bdc-eb83-4926-a369-8a61d912b308/volumes" Nov 24 10:31:10 crc kubenswrapper[4944]: I1124 10:31:10.313473 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab617caa-207c-48b4-a541-32b97f4a7a53" path="/var/lib/kubelet/pods/ab617caa-207c-48b4-a541-32b97f4a7a53/volumes" Nov 24 10:31:13 crc kubenswrapper[4944]: I1124 10:31:13.186836 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" event={"ID":"b14f0425-11fb-4ed0-9599-0c08bb518a85","Type":"ContainerStarted","Data":"e85cc0eaf73980379dcf6ec9e79dc4144488657e2315db1d5aa621f53391d69e"} Nov 24 10:31:13 crc kubenswrapper[4944]: I1124 10:31:13.188147 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" Nov 24 10:31:13 crc kubenswrapper[4944]: I1124 10:31:13.191458 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" Nov 24 10:31:13 crc kubenswrapper[4944]: I1124 10:31:13.214339 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-vrrsh" podStartSLOduration=1.865120438 podStartE2EDuration="11.214311755s" podCreationTimestamp="2025-11-24 10:31:02 +0000 UTC" firstStartedPulling="2025-11-24 10:31:03.366984632 +0000 UTC m=+5923.901425094" lastFinishedPulling="2025-11-24 10:31:12.716175949 +0000 UTC m=+5933.250616411" observedRunningTime="2025-11-24 10:31:13.20507126 +0000 UTC m=+5933.739511732" watchObservedRunningTime="2025-11-24 10:31:13.214311755 +0000 UTC m=+5933.748752217" Nov 24 10:31:16 crc kubenswrapper[4944]: I1124 10:31:16.034028 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-p2p5g"] Nov 24 10:31:16 crc kubenswrapper[4944]: I1124 10:31:16.043483 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-p2p5g"] Nov 24 10:31:16 crc kubenswrapper[4944]: I1124 10:31:16.294627 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e324380-8997-48e9-acb6-b7c22e9eecba" path="/var/lib/kubelet/pods/5e324380-8997-48e9-acb6-b7c22e9eecba/volumes" Nov 24 10:31:20 crc kubenswrapper[4944]: I1124 10:31:20.291331 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:31:20 crc kubenswrapper[4944]: E1124 10:31:20.292207 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:31:22 crc kubenswrapper[4944]: I1124 10:31:22.640334 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-t7fc7" Nov 24 10:31:24 crc kubenswrapper[4944]: I1124 10:31:24.912778 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 10:31:24 crc kubenswrapper[4944]: I1124 10:31:24.913029 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="9e4a6ace-6f5e-4c0e-94e8-82040b15edad" containerName="openstackclient" containerID="cri-o://c5a832c4d86aa8f2148f7f59aa26c773201f153873deedab1329c0e606be14bf" gracePeriod=2 Nov 24 10:31:24 crc kubenswrapper[4944]: I1124 10:31:24.932265 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.035125 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 10:31:25 crc kubenswrapper[4944]: E1124 10:31:25.036018 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4a6ace-6f5e-4c0e-94e8-82040b15edad" containerName="openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.036036 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4a6ace-6f5e-4c0e-94e8-82040b15edad" containerName="openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.036287 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4a6ace-6f5e-4c0e-94e8-82040b15edad" containerName="openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.037120 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.079615 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.148380 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 10:31:25 crc kubenswrapper[4944]: E1124 10:31:25.149460 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-hrbd9 openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="e645f836-57ae-49d7-be89-9aa8742651cb" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.157120 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.161138 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config-secret\") pod \"openstackclient\" (UID: \"e645f836-57ae-49d7-be89-9aa8742651cb\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.168772 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config\") pod \"openstackclient\" (UID: \"e645f836-57ae-49d7-be89-9aa8742651cb\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.169093 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrbd9\" (UniqueName: \"kubernetes.io/projected/e645f836-57ae-49d7-be89-9aa8742651cb-kube-api-access-hrbd9\") pod \"openstackclient\" (UID: \"e645f836-57ae-49d7-be89-9aa8742651cb\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.178927 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.180659 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.196277 4944 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="9e4a6ace-6f5e-4c0e-94e8-82040b15edad" podUID="3602dfbe-14d8-4dca-a2f7-3f645acc52ad" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.211818 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.241710 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.258804 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.269496 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-4wkxm" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.270909 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrbd9\" (UniqueName: \"kubernetes.io/projected/e645f836-57ae-49d7-be89-9aa8742651cb-kube-api-access-hrbd9\") pod \"openstackclient\" (UID: \"e645f836-57ae-49d7-be89-9aa8742651cb\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.271081 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3602dfbe-14d8-4dca-a2f7-3f645acc52ad-openstack-config\") pod \"openstackclient\" (UID: \"3602dfbe-14d8-4dca-a2f7-3f645acc52ad\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.271265 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config-secret\") pod \"openstackclient\" (UID: \"e645f836-57ae-49d7-be89-9aa8742651cb\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.271366 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3602dfbe-14d8-4dca-a2f7-3f645acc52ad-openstack-config-secret\") pod \"openstackclient\" (UID: \"3602dfbe-14d8-4dca-a2f7-3f645acc52ad\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.271531 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh9jb\" (UniqueName: \"kubernetes.io/projected/3602dfbe-14d8-4dca-a2f7-3f645acc52ad-kube-api-access-kh9jb\") pod \"openstackclient\" (UID: \"3602dfbe-14d8-4dca-a2f7-3f645acc52ad\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.271636 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config\") pod \"openstackclient\" (UID: \"e645f836-57ae-49d7-be89-9aa8742651cb\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: E1124 10:31:25.289678 4944 projected.go:194] Error preparing data for projected volume kube-api-access-hrbd9 for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (e645f836-57ae-49d7-be89-9aa8742651cb) does not match the UID in record. The object might have been deleted and then recreated Nov 24 10:31:25 crc kubenswrapper[4944]: E1124 10:31:25.289756 4944 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e645f836-57ae-49d7-be89-9aa8742651cb-kube-api-access-hrbd9 podName:e645f836-57ae-49d7-be89-9aa8742651cb nodeName:}" failed. No retries permitted until 2025-11-24 10:31:25.789737583 +0000 UTC m=+5946.324178045 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hrbd9" (UniqueName: "kubernetes.io/projected/e645f836-57ae-49d7-be89-9aa8742651cb-kube-api-access-hrbd9") pod "openstackclient" (UID: "e645f836-57ae-49d7-be89-9aa8742651cb") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (e645f836-57ae-49d7-be89-9aa8742651cb) does not match the UID in record. The object might have been deleted and then recreated Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.290120 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config\") pod \"openstackclient\" (UID: \"e645f836-57ae-49d7-be89-9aa8742651cb\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.313302 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config-secret\") pod \"openstackclient\" (UID: \"e645f836-57ae-49d7-be89-9aa8742651cb\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.341939 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.358395 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.373924 4944 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e645f836-57ae-49d7-be89-9aa8742651cb" podUID="3602dfbe-14d8-4dca-a2f7-3f645acc52ad" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.375139 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3602dfbe-14d8-4dca-a2f7-3f645acc52ad-openstack-config\") pod \"openstackclient\" (UID: \"3602dfbe-14d8-4dca-a2f7-3f645acc52ad\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.375296 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3602dfbe-14d8-4dca-a2f7-3f645acc52ad-openstack-config-secret\") pod \"openstackclient\" (UID: \"3602dfbe-14d8-4dca-a2f7-3f645acc52ad\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.375390 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh9jb\" (UniqueName: \"kubernetes.io/projected/3602dfbe-14d8-4dca-a2f7-3f645acc52ad-kube-api-access-kh9jb\") pod \"openstackclient\" (UID: \"3602dfbe-14d8-4dca-a2f7-3f645acc52ad\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.375435 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb42q\" (UniqueName: \"kubernetes.io/projected/afd374c3-c2ec-4c15-a821-9223bd6e7e2d-kube-api-access-xb42q\") pod \"kube-state-metrics-0\" (UID: \"afd374c3-c2ec-4c15-a821-9223bd6e7e2d\") " pod="openstack/kube-state-metrics-0" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.376348 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3602dfbe-14d8-4dca-a2f7-3f645acc52ad-openstack-config\") pod \"openstackclient\" (UID: \"3602dfbe-14d8-4dca-a2f7-3f645acc52ad\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.385544 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.428636 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh9jb\" (UniqueName: \"kubernetes.io/projected/3602dfbe-14d8-4dca-a2f7-3f645acc52ad-kube-api-access-kh9jb\") pod \"openstackclient\" (UID: \"3602dfbe-14d8-4dca-a2f7-3f645acc52ad\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.433678 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3602dfbe-14d8-4dca-a2f7-3f645acc52ad-openstack-config-secret\") pod \"openstackclient\" (UID: \"3602dfbe-14d8-4dca-a2f7-3f645acc52ad\") " pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.479665 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config-secret\") pod \"e645f836-57ae-49d7-be89-9aa8742651cb\" (UID: \"e645f836-57ae-49d7-be89-9aa8742651cb\") " Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.479968 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config\") pod \"e645f836-57ae-49d7-be89-9aa8742651cb\" (UID: \"e645f836-57ae-49d7-be89-9aa8742651cb\") " Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.480375 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb42q\" (UniqueName: \"kubernetes.io/projected/afd374c3-c2ec-4c15-a821-9223bd6e7e2d-kube-api-access-xb42q\") pod \"kube-state-metrics-0\" (UID: \"afd374c3-c2ec-4c15-a821-9223bd6e7e2d\") " pod="openstack/kube-state-metrics-0" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.480570 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrbd9\" (UniqueName: \"kubernetes.io/projected/e645f836-57ae-49d7-be89-9aa8742651cb-kube-api-access-hrbd9\") on node \"crc\" DevicePath \"\"" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.481834 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "e645f836-57ae-49d7-be89-9aa8742651cb" (UID: "e645f836-57ae-49d7-be89-9aa8742651cb"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.493401 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "e645f836-57ae-49d7-be89-9aa8742651cb" (UID: "e645f836-57ae-49d7-be89-9aa8742651cb"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.516025 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb42q\" (UniqueName: \"kubernetes.io/projected/afd374c3-c2ec-4c15-a821-9223bd6e7e2d-kube-api-access-xb42q\") pod \"kube-state-metrics-0\" (UID: \"afd374c3-c2ec-4c15-a821-9223bd6e7e2d\") " pod="openstack/kube-state-metrics-0" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.527852 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.612208 4944 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.612247 4944 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e645f836-57ae-49d7-be89-9aa8742651cb-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 10:31:25 crc kubenswrapper[4944]: I1124 10:31:25.658602 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.094423 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.111758 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.129308 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.129417 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.129508 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.129693 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.129789 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-kpq9b" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.171364 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.248740 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/be3827d1-c5b1-4048-af2f-e2a2be15a396-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.248828 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9p8j\" (UniqueName: \"kubernetes.io/projected/be3827d1-c5b1-4048-af2f-e2a2be15a396-kube-api-access-r9p8j\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.248892 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/be3827d1-c5b1-4048-af2f-e2a2be15a396-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.248914 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/be3827d1-c5b1-4048-af2f-e2a2be15a396-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.248957 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/be3827d1-c5b1-4048-af2f-e2a2be15a396-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.249031 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/be3827d1-c5b1-4048-af2f-e2a2be15a396-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.249093 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/be3827d1-c5b1-4048-af2f-e2a2be15a396-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.350432 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/be3827d1-c5b1-4048-af2f-e2a2be15a396-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.350510 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/be3827d1-c5b1-4048-af2f-e2a2be15a396-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.350567 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9p8j\" (UniqueName: \"kubernetes.io/projected/be3827d1-c5b1-4048-af2f-e2a2be15a396-kube-api-access-r9p8j\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.350647 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/be3827d1-c5b1-4048-af2f-e2a2be15a396-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.350676 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/be3827d1-c5b1-4048-af2f-e2a2be15a396-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.350728 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/be3827d1-c5b1-4048-af2f-e2a2be15a396-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.350798 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/be3827d1-c5b1-4048-af2f-e2a2be15a396-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.355310 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/be3827d1-c5b1-4048-af2f-e2a2be15a396-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.379526 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/be3827d1-c5b1-4048-af2f-e2a2be15a396-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.381143 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/be3827d1-c5b1-4048-af2f-e2a2be15a396-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.385554 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/be3827d1-c5b1-4048-af2f-e2a2be15a396-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.385827 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/be3827d1-c5b1-4048-af2f-e2a2be15a396-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.390498 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/be3827d1-c5b1-4048-af2f-e2a2be15a396-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.406097 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e645f836-57ae-49d7-be89-9aa8742651cb" path="/var/lib/kubelet/pods/e645f836-57ae-49d7-be89-9aa8742651cb/volumes" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.442561 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9p8j\" (UniqueName: \"kubernetes.io/projected/be3827d1-c5b1-4048-af2f-e2a2be15a396-kube-api-access-r9p8j\") pod \"alertmanager-metric-storage-0\" (UID: \"be3827d1-c5b1-4048-af2f-e2a2be15a396\") " pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.456339 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.493712 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.551796 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.568339 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.579722 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.579916 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.579977 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.580109 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.580322 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-sr5mv" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.580459 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.610221 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.651702 4944 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e645f836-57ae-49d7-be89-9aa8742651cb" podUID="3602dfbe-14d8-4dca-a2f7-3f645acc52ad" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.704572 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.705017 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.705066 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.705109 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.705158 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c7beffb0-b909-40ee-a906-b977a5ffa2b7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c7beffb0-b909-40ee-a906-b977a5ffa2b7\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.705184 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27fv8\" (UniqueName: \"kubernetes.io/projected/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-kube-api-access-27fv8\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.705243 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.705284 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-config\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.807693 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.807786 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.807813 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.807854 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.807906 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c7beffb0-b909-40ee-a906-b977a5ffa2b7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c7beffb0-b909-40ee-a906-b977a5ffa2b7\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.807932 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27fv8\" (UniqueName: \"kubernetes.io/projected/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-kube-api-access-27fv8\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.807985 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.808027 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-config\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.814355 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.816730 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.817890 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-config\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.821706 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.836484 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.837552 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.840060 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27fv8\" (UniqueName: \"kubernetes.io/projected/230aefa5-9e7d-4c8f-8d3e-dd7349413f28-kube-api-access-27fv8\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.845643 4944 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 10:31:26 crc kubenswrapper[4944]: I1124 10:31:26.845706 4944 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c7beffb0-b909-40ee-a906-b977a5ffa2b7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c7beffb0-b909-40ee-a906-b977a5ffa2b7\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b1c30636436ac0b23cdffb33b2a1feb59bc05dfb9d2a11ca9f32f0db795dc129/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.046129 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c7beffb0-b909-40ee-a906-b977a5ffa2b7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c7beffb0-b909-40ee-a906-b977a5ffa2b7\") pod \"prometheus-metric-storage-0\" (UID: \"230aefa5-9e7d-4c8f-8d3e-dd7349413f28\") " pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.052825 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.099372 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.259466 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.456436 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.513264 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"be3827d1-c5b1-4048-af2f-e2a2be15a396","Type":"ContainerStarted","Data":"f9bf3f9814eda8a8765edc0bcb7b2a763ff469bb7a8e45cd831278e23cf2c149"} Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.517903 4944 generic.go:334] "Generic (PLEG): container finished" podID="9e4a6ace-6f5e-4c0e-94e8-82040b15edad" containerID="c5a832c4d86aa8f2148f7f59aa26c773201f153873deedab1329c0e606be14bf" exitCode=137 Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.521760 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3602dfbe-14d8-4dca-a2f7-3f645acc52ad","Type":"ContainerStarted","Data":"9b65f10c3cc6d2b67cd127b098f38a82731c809ee52c8258e382518182090c1b"} Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.530755 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"afd374c3-c2ec-4c15-a821-9223bd6e7e2d","Type":"ContainerStarted","Data":"9ccc5fb412a5da8c8ad107b6975daa65f3655bacd2f503eadd92b4fbfe14fbad"} Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.753074 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.851924 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m85wt\" (UniqueName: \"kubernetes.io/projected/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-kube-api-access-m85wt\") pod \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.852097 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config\") pod \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.852163 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config-secret\") pod \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\" (UID: \"9e4a6ace-6f5e-4c0e-94e8-82040b15edad\") " Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.871278 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-kube-api-access-m85wt" (OuterVolumeSpecName: "kube-api-access-m85wt") pod "9e4a6ace-6f5e-4c0e-94e8-82040b15edad" (UID: "9e4a6ace-6f5e-4c0e-94e8-82040b15edad"). InnerVolumeSpecName "kube-api-access-m85wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.890613 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "9e4a6ace-6f5e-4c0e-94e8-82040b15edad" (UID: "9e4a6ace-6f5e-4c0e-94e8-82040b15edad"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.967369 4944 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.967416 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m85wt\" (UniqueName: \"kubernetes.io/projected/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-kube-api-access-m85wt\") on node \"crc\" DevicePath \"\"" Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.969626 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "9e4a6ace-6f5e-4c0e-94e8-82040b15edad" (UID: "9e4a6ace-6f5e-4c0e-94e8-82040b15edad"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:31:27 crc kubenswrapper[4944]: I1124 10:31:27.986983 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 24 10:31:28 crc kubenswrapper[4944]: I1124 10:31:28.070346 4944 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9e4a6ace-6f5e-4c0e-94e8-82040b15edad-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 10:31:28 crc kubenswrapper[4944]: I1124 10:31:28.289509 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e4a6ace-6f5e-4c0e-94e8-82040b15edad" path="/var/lib/kubelet/pods/9e4a6ace-6f5e-4c0e-94e8-82040b15edad/volumes" Nov 24 10:31:28 crc kubenswrapper[4944]: I1124 10:31:28.545061 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3602dfbe-14d8-4dca-a2f7-3f645acc52ad","Type":"ContainerStarted","Data":"a37e03565dcf315ff8b1a0e0b2de857ca6978d3e82870bf090f3ad7e441ac5b0"} Nov 24 10:31:28 crc kubenswrapper[4944]: I1124 10:31:28.549893 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"afd374c3-c2ec-4c15-a821-9223bd6e7e2d","Type":"ContainerStarted","Data":"31315faf05a36a936cdc31f3f16ff6bea3def3c25e341cd0cec806a76b8d34b7"} Nov 24 10:31:28 crc kubenswrapper[4944]: I1124 10:31:28.550038 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 10:31:28 crc kubenswrapper[4944]: I1124 10:31:28.552254 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"230aefa5-9e7d-4c8f-8d3e-dd7349413f28","Type":"ContainerStarted","Data":"b38117ea607bdf1236de4e4150a53858f8bc9a81e6b900ac4933886f44e63a56"} Nov 24 10:31:28 crc kubenswrapper[4944]: I1124 10:31:28.555236 4944 scope.go:117] "RemoveContainer" containerID="c5a832c4d86aa8f2148f7f59aa26c773201f153873deedab1329c0e606be14bf" Nov 24 10:31:28 crc kubenswrapper[4944]: I1124 10:31:28.555600 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 10:31:28 crc kubenswrapper[4944]: I1124 10:31:28.565875 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.565850775 podStartE2EDuration="3.565850775s" podCreationTimestamp="2025-11-24 10:31:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:31:28.562440047 +0000 UTC m=+5949.096880529" watchObservedRunningTime="2025-11-24 10:31:28.565850775 +0000 UTC m=+5949.100291257" Nov 24 10:31:28 crc kubenswrapper[4944]: I1124 10:31:28.604134 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.982394953 podStartE2EDuration="3.604109894s" podCreationTimestamp="2025-11-24 10:31:25 +0000 UTC" firstStartedPulling="2025-11-24 10:31:27.068192726 +0000 UTC m=+5947.602633188" lastFinishedPulling="2025-11-24 10:31:27.689907667 +0000 UTC m=+5948.224348129" observedRunningTime="2025-11-24 10:31:28.578932873 +0000 UTC m=+5949.113373345" watchObservedRunningTime="2025-11-24 10:31:28.604109894 +0000 UTC m=+5949.138550346" Nov 24 10:31:31 crc kubenswrapper[4944]: I1124 10:31:31.277642 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:31:31 crc kubenswrapper[4944]: E1124 10:31:31.278629 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:31:32 crc kubenswrapper[4944]: I1124 10:31:32.275059 4944 scope.go:117] "RemoveContainer" containerID="fdb8f570d4b4ecc9cc9437e14ad556388e848c47a804bde425a413d1a138d1af" Nov 24 10:31:32 crc kubenswrapper[4944]: I1124 10:31:32.418900 4944 scope.go:117] "RemoveContainer" containerID="7d243a0e8a59fd759d977f83018e79dd7a87fc7252e628966e29c4ea33148fff" Nov 24 10:31:32 crc kubenswrapper[4944]: I1124 10:31:32.880364 4944 scope.go:117] "RemoveContainer" containerID="2f9d6d7b6fdf1e8bd9d917412afa0d1ed283f62c5bb5d21ac5fde4828ada893c" Nov 24 10:31:32 crc kubenswrapper[4944]: I1124 10:31:32.924895 4944 scope.go:117] "RemoveContainer" containerID="d3dfac5e9eeb53b094a68347590f5d09a843ef8f8cf0d8cfeae53e2b7a56d184" Nov 24 10:31:32 crc kubenswrapper[4944]: I1124 10:31:32.973971 4944 scope.go:117] "RemoveContainer" containerID="4cf81980ba0068728aa8849f6f4ca50c9c51f2a391bd84de93b6abde90eefc60" Nov 24 10:31:33 crc kubenswrapper[4944]: I1124 10:31:33.023352 4944 scope.go:117] "RemoveContainer" containerID="1e3fb39ec669308fd2146d8b1b299514c212cc268a17629dad96dd8bd9616433" Nov 24 10:31:33 crc kubenswrapper[4944]: I1124 10:31:33.728623 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"230aefa5-9e7d-4c8f-8d3e-dd7349413f28","Type":"ContainerStarted","Data":"8d5cf71df1717769c15a4917d38d6b23b5375fb7e9aa617cb6748e51cb37c627"} Nov 24 10:31:34 crc kubenswrapper[4944]: I1124 10:31:34.742132 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"be3827d1-c5b1-4048-af2f-e2a2be15a396","Type":"ContainerStarted","Data":"a292a8f1ce678913a6fcbe8b72b1d30ce7f70cdc6b3e5e715afac75edad949c1"} Nov 24 10:31:35 crc kubenswrapper[4944]: I1124 10:31:35.662783 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.436837 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v6wvg"] Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.441324 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.452773 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v6wvg"] Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.581129 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-catalog-content\") pod \"redhat-operators-v6wvg\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.581514 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdpht\" (UniqueName: \"kubernetes.io/projected/bc0414f1-24b8-422c-80d8-292154d20bd4-kube-api-access-qdpht\") pod \"redhat-operators-v6wvg\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.581565 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-utilities\") pod \"redhat-operators-v6wvg\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.683763 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-catalog-content\") pod \"redhat-operators-v6wvg\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.683854 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdpht\" (UniqueName: \"kubernetes.io/projected/bc0414f1-24b8-422c-80d8-292154d20bd4-kube-api-access-qdpht\") pod \"redhat-operators-v6wvg\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.683917 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-utilities\") pod \"redhat-operators-v6wvg\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.684372 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-catalog-content\") pod \"redhat-operators-v6wvg\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.684476 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-utilities\") pod \"redhat-operators-v6wvg\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.703526 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdpht\" (UniqueName: \"kubernetes.io/projected/bc0414f1-24b8-422c-80d8-292154d20bd4-kube-api-access-qdpht\") pod \"redhat-operators-v6wvg\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.770976 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.785589 4944 generic.go:334] "Generic (PLEG): container finished" podID="230aefa5-9e7d-4c8f-8d3e-dd7349413f28" containerID="8d5cf71df1717769c15a4917d38d6b23b5375fb7e9aa617cb6748e51cb37c627" exitCode=0 Nov 24 10:31:38 crc kubenswrapper[4944]: I1124 10:31:38.785682 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"230aefa5-9e7d-4c8f-8d3e-dd7349413f28","Type":"ContainerDied","Data":"8d5cf71df1717769c15a4917d38d6b23b5375fb7e9aa617cb6748e51cb37c627"} Nov 24 10:31:39 crc kubenswrapper[4944]: I1124 10:31:39.272247 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v6wvg"] Nov 24 10:31:39 crc kubenswrapper[4944]: I1124 10:31:39.794804 4944 generic.go:334] "Generic (PLEG): container finished" podID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerID="b80600d3875f913e40907812bb9e3d19a820c2df0e5b95e85599dfc5dbec5a6f" exitCode=0 Nov 24 10:31:39 crc kubenswrapper[4944]: I1124 10:31:39.794880 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6wvg" event={"ID":"bc0414f1-24b8-422c-80d8-292154d20bd4","Type":"ContainerDied","Data":"b80600d3875f913e40907812bb9e3d19a820c2df0e5b95e85599dfc5dbec5a6f"} Nov 24 10:31:39 crc kubenswrapper[4944]: I1124 10:31:39.795176 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6wvg" event={"ID":"bc0414f1-24b8-422c-80d8-292154d20bd4","Type":"ContainerStarted","Data":"0cab3b832edbba13f28ef44844b84b7add3c59c2b7ad416b0b241ff47583a81e"} Nov 24 10:31:40 crc kubenswrapper[4944]: I1124 10:31:40.808303 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6wvg" event={"ID":"bc0414f1-24b8-422c-80d8-292154d20bd4","Type":"ContainerStarted","Data":"e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726"} Nov 24 10:31:40 crc kubenswrapper[4944]: I1124 10:31:40.809788 4944 generic.go:334] "Generic (PLEG): container finished" podID="be3827d1-c5b1-4048-af2f-e2a2be15a396" containerID="a292a8f1ce678913a6fcbe8b72b1d30ce7f70cdc6b3e5e715afac75edad949c1" exitCode=0 Nov 24 10:31:40 crc kubenswrapper[4944]: I1124 10:31:40.809865 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"be3827d1-c5b1-4048-af2f-e2a2be15a396","Type":"ContainerDied","Data":"a292a8f1ce678913a6fcbe8b72b1d30ce7f70cdc6b3e5e715afac75edad949c1"} Nov 24 10:31:44 crc kubenswrapper[4944]: I1124 10:31:44.276865 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:31:44 crc kubenswrapper[4944]: E1124 10:31:44.277569 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:31:44 crc kubenswrapper[4944]: I1124 10:31:44.856026 4944 generic.go:334] "Generic (PLEG): container finished" podID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerID="e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726" exitCode=0 Nov 24 10:31:44 crc kubenswrapper[4944]: I1124 10:31:44.856083 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6wvg" event={"ID":"bc0414f1-24b8-422c-80d8-292154d20bd4","Type":"ContainerDied","Data":"e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726"} Nov 24 10:31:46 crc kubenswrapper[4944]: I1124 10:31:46.880181 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6wvg" event={"ID":"bc0414f1-24b8-422c-80d8-292154d20bd4","Type":"ContainerStarted","Data":"c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a"} Nov 24 10:31:46 crc kubenswrapper[4944]: I1124 10:31:46.882407 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"be3827d1-c5b1-4048-af2f-e2a2be15a396","Type":"ContainerStarted","Data":"276909bebd53240a50983868dcb36ca220a26ae0dde806b0e8827bd95cf1b8f0"} Nov 24 10:31:46 crc kubenswrapper[4944]: I1124 10:31:46.884521 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"230aefa5-9e7d-4c8f-8d3e-dd7349413f28","Type":"ContainerStarted","Data":"8ee9385ac108d1c0a3e8ac9d0f2e7665429e29a6ed6c5fef699cde865afcdf0d"} Nov 24 10:31:46 crc kubenswrapper[4944]: I1124 10:31:46.898629 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v6wvg" podStartSLOduration=2.687467285 podStartE2EDuration="8.898604106s" podCreationTimestamp="2025-11-24 10:31:38 +0000 UTC" firstStartedPulling="2025-11-24 10:31:39.796443225 +0000 UTC m=+5960.330883687" lastFinishedPulling="2025-11-24 10:31:46.007580046 +0000 UTC m=+5966.542020508" observedRunningTime="2025-11-24 10:31:46.895832457 +0000 UTC m=+5967.430272919" watchObservedRunningTime="2025-11-24 10:31:46.898604106 +0000 UTC m=+5967.433044568" Nov 24 10:31:48 crc kubenswrapper[4944]: I1124 10:31:48.771419 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:48 crc kubenswrapper[4944]: I1124 10:31:48.771744 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:49 crc kubenswrapper[4944]: I1124 10:31:49.912955 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"230aefa5-9e7d-4c8f-8d3e-dd7349413f28","Type":"ContainerStarted","Data":"a979f704dc8455f50e7696b21b369ee98a64873302e5d12793978b8cdb0e9c35"} Nov 24 10:31:49 crc kubenswrapper[4944]: I1124 10:31:49.915890 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"be3827d1-c5b1-4048-af2f-e2a2be15a396","Type":"ContainerStarted","Data":"9ca41c7dadd0028bc723543537ed2413eb574e60386528be858a484645455435"} Nov 24 10:31:49 crc kubenswrapper[4944]: I1124 10:31:49.916106 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:49 crc kubenswrapper[4944]: I1124 10:31:49.918240 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Nov 24 10:31:49 crc kubenswrapper[4944]: I1124 10:31:49.950346 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=5.396284985 podStartE2EDuration="23.950319672s" podCreationTimestamp="2025-11-24 10:31:26 +0000 UTC" firstStartedPulling="2025-11-24 10:31:27.453532529 +0000 UTC m=+5947.987972991" lastFinishedPulling="2025-11-24 10:31:46.007567226 +0000 UTC m=+5966.542007678" observedRunningTime="2025-11-24 10:31:49.937767451 +0000 UTC m=+5970.472207933" watchObservedRunningTime="2025-11-24 10:31:49.950319672 +0000 UTC m=+5970.484760134" Nov 24 10:31:50 crc kubenswrapper[4944]: I1124 10:31:50.138600 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v6wvg" podUID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerName="registry-server" probeResult="failure" output=< Nov 24 10:31:50 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 10:31:50 crc kubenswrapper[4944]: > Nov 24 10:31:53 crc kubenswrapper[4944]: I1124 10:31:53.953653 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"230aefa5-9e7d-4c8f-8d3e-dd7349413f28","Type":"ContainerStarted","Data":"61499af0e99a5c45445f40918fa22238d1df2900c3a9321e6d4cc09db25101ee"} Nov 24 10:31:53 crc kubenswrapper[4944]: I1124 10:31:53.987829 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.807469511 podStartE2EDuration="28.987804424s" podCreationTimestamp="2025-11-24 10:31:25 +0000 UTC" firstStartedPulling="2025-11-24 10:31:28.009105753 +0000 UTC m=+5948.543546205" lastFinishedPulling="2025-11-24 10:31:53.189440656 +0000 UTC m=+5973.723881118" observedRunningTime="2025-11-24 10:31:53.98081016 +0000 UTC m=+5974.515250632" watchObservedRunningTime="2025-11-24 10:31:53.987804424 +0000 UTC m=+5974.522244886" Nov 24 10:31:57 crc kubenswrapper[4944]: I1124 10:31:57.260937 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:57 crc kubenswrapper[4944]: I1124 10:31:57.261689 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:57 crc kubenswrapper[4944]: I1124 10:31:57.264222 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:57 crc kubenswrapper[4944]: I1124 10:31:57.277425 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:31:57 crc kubenswrapper[4944]: I1124 10:31:57.991819 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"6c665bc43080f7d77b877e010e51181985aff09e212dee7ac1010f5e6186cdd1"} Nov 24 10:31:57 crc kubenswrapper[4944]: I1124 10:31:57.994904 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 24 10:31:58 crc kubenswrapper[4944]: I1124 10:31:58.830295 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:58 crc kubenswrapper[4944]: I1124 10:31:58.887014 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.067853 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v6wvg"] Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.147118 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.152339 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.157465 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.157866 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.158989 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.265953 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-run-httpd\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.266007 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjznp\" (UniqueName: \"kubernetes.io/projected/2d4325ae-9f5e-4dba-836d-836058a44a6d-kube-api-access-vjznp\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.266153 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.266440 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-config-data\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.266508 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-log-httpd\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.266572 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-scripts\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.266612 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.368477 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-config-data\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.368823 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-log-httpd\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.368976 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-scripts\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.369157 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.369368 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-run-httpd\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.369521 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjznp\" (UniqueName: \"kubernetes.io/projected/2d4325ae-9f5e-4dba-836d-836058a44a6d-kube-api-access-vjznp\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.369664 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.370567 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-log-httpd\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.371988 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-run-httpd\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.376309 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.377608 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-scripts\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.378334 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.382711 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-config-data\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.398958 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjznp\" (UniqueName: \"kubernetes.io/projected/2d4325ae-9f5e-4dba-836d-836058a44a6d-kube-api-access-vjznp\") pod \"ceilometer-0\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " pod="openstack/ceilometer-0" Nov 24 10:31:59 crc kubenswrapper[4944]: I1124 10:31:59.481306 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.017248 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v6wvg" podUID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerName="registry-server" containerID="cri-o://c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a" gracePeriod=2 Nov 24 10:32:00 crc kubenswrapper[4944]: W1124 10:32:00.054697 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d4325ae_9f5e_4dba_836d_836058a44a6d.slice/crio-67d1b13cbd405eec579abcc5af453604eceb7a49a10bb8d78d04cc5458cecc53 WatchSource:0}: Error finding container 67d1b13cbd405eec579abcc5af453604eceb7a49a10bb8d78d04cc5458cecc53: Status 404 returned error can't find the container with id 67d1b13cbd405eec579abcc5af453604eceb7a49a10bb8d78d04cc5458cecc53 Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.057451 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.573892 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.708583 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-catalog-content\") pod \"bc0414f1-24b8-422c-80d8-292154d20bd4\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.708666 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdpht\" (UniqueName: \"kubernetes.io/projected/bc0414f1-24b8-422c-80d8-292154d20bd4-kube-api-access-qdpht\") pod \"bc0414f1-24b8-422c-80d8-292154d20bd4\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.708741 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-utilities\") pod \"bc0414f1-24b8-422c-80d8-292154d20bd4\" (UID: \"bc0414f1-24b8-422c-80d8-292154d20bd4\") " Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.718476 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc0414f1-24b8-422c-80d8-292154d20bd4-kube-api-access-qdpht" (OuterVolumeSpecName: "kube-api-access-qdpht") pod "bc0414f1-24b8-422c-80d8-292154d20bd4" (UID: "bc0414f1-24b8-422c-80d8-292154d20bd4"). InnerVolumeSpecName "kube-api-access-qdpht". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.719719 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-utilities" (OuterVolumeSpecName: "utilities") pod "bc0414f1-24b8-422c-80d8-292154d20bd4" (UID: "bc0414f1-24b8-422c-80d8-292154d20bd4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.810559 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.810599 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdpht\" (UniqueName: \"kubernetes.io/projected/bc0414f1-24b8-422c-80d8-292154d20bd4-kube-api-access-qdpht\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.820591 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc0414f1-24b8-422c-80d8-292154d20bd4" (UID: "bc0414f1-24b8-422c-80d8-292154d20bd4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:32:00 crc kubenswrapper[4944]: I1124 10:32:00.913008 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc0414f1-24b8-422c-80d8-292154d20bd4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.027704 4944 generic.go:334] "Generic (PLEG): container finished" podID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerID="c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a" exitCode=0 Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.027794 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6wvg" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.027816 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6wvg" event={"ID":"bc0414f1-24b8-422c-80d8-292154d20bd4","Type":"ContainerDied","Data":"c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a"} Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.029370 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6wvg" event={"ID":"bc0414f1-24b8-422c-80d8-292154d20bd4","Type":"ContainerDied","Data":"0cab3b832edbba13f28ef44844b84b7add3c59c2b7ad416b0b241ff47583a81e"} Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.029401 4944 scope.go:117] "RemoveContainer" containerID="c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.035117 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4325ae-9f5e-4dba-836d-836058a44a6d","Type":"ContainerStarted","Data":"543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9"} Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.035171 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4325ae-9f5e-4dba-836d-836058a44a6d","Type":"ContainerStarted","Data":"67d1b13cbd405eec579abcc5af453604eceb7a49a10bb8d78d04cc5458cecc53"} Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.053212 4944 scope.go:117] "RemoveContainer" containerID="e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.073589 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v6wvg"] Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.085497 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v6wvg"] Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.086718 4944 scope.go:117] "RemoveContainer" containerID="b80600d3875f913e40907812bb9e3d19a820c2df0e5b95e85599dfc5dbec5a6f" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.130121 4944 scope.go:117] "RemoveContainer" containerID="c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a" Nov 24 10:32:01 crc kubenswrapper[4944]: E1124 10:32:01.130608 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a\": container with ID starting with c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a not found: ID does not exist" containerID="c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.130715 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a"} err="failed to get container status \"c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a\": rpc error: code = NotFound desc = could not find container \"c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a\": container with ID starting with c0e8a976ccce3a62d5bfacef167eeed580428900e69e4e443ea6675c52502c4a not found: ID does not exist" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.130834 4944 scope.go:117] "RemoveContainer" containerID="e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726" Nov 24 10:32:01 crc kubenswrapper[4944]: E1124 10:32:01.131433 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726\": container with ID starting with e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726 not found: ID does not exist" containerID="e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.131455 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726"} err="failed to get container status \"e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726\": rpc error: code = NotFound desc = could not find container \"e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726\": container with ID starting with e5fdd461d854cfc505734ce931bb3a04868d4a86b697207c0e80c4060a1c3726 not found: ID does not exist" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.131470 4944 scope.go:117] "RemoveContainer" containerID="b80600d3875f913e40907812bb9e3d19a820c2df0e5b95e85599dfc5dbec5a6f" Nov 24 10:32:01 crc kubenswrapper[4944]: E1124 10:32:01.131805 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b80600d3875f913e40907812bb9e3d19a820c2df0e5b95e85599dfc5dbec5a6f\": container with ID starting with b80600d3875f913e40907812bb9e3d19a820c2df0e5b95e85599dfc5dbec5a6f not found: ID does not exist" containerID="b80600d3875f913e40907812bb9e3d19a820c2df0e5b95e85599dfc5dbec5a6f" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.131829 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b80600d3875f913e40907812bb9e3d19a820c2df0e5b95e85599dfc5dbec5a6f"} err="failed to get container status \"b80600d3875f913e40907812bb9e3d19a820c2df0e5b95e85599dfc5dbec5a6f\": rpc error: code = NotFound desc = could not find container \"b80600d3875f913e40907812bb9e3d19a820c2df0e5b95e85599dfc5dbec5a6f\": container with ID starting with b80600d3875f913e40907812bb9e3d19a820c2df0e5b95e85599dfc5dbec5a6f not found: ID does not exist" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.883816 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q27r5"] Nov 24 10:32:01 crc kubenswrapper[4944]: E1124 10:32:01.884933 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerName="registry-server" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.884959 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerName="registry-server" Nov 24 10:32:01 crc kubenswrapper[4944]: E1124 10:32:01.884975 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerName="extract-utilities" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.884984 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerName="extract-utilities" Nov 24 10:32:01 crc kubenswrapper[4944]: E1124 10:32:01.885007 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerName="extract-content" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.885015 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerName="extract-content" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.885442 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc0414f1-24b8-422c-80d8-292154d20bd4" containerName="registry-server" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.893004 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.909486 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q27r5"] Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.936665 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-catalog-content\") pod \"certified-operators-q27r5\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.936726 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-utilities\") pod \"certified-operators-q27r5\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:01 crc kubenswrapper[4944]: I1124 10:32:01.936780 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgqzm\" (UniqueName: \"kubernetes.io/projected/8ffb82bd-50cd-4935-aa29-254b77054491-kube-api-access-bgqzm\") pod \"certified-operators-q27r5\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:02 crc kubenswrapper[4944]: I1124 10:32:02.038933 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-utilities\") pod \"certified-operators-q27r5\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:02 crc kubenswrapper[4944]: I1124 10:32:02.039102 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgqzm\" (UniqueName: \"kubernetes.io/projected/8ffb82bd-50cd-4935-aa29-254b77054491-kube-api-access-bgqzm\") pod \"certified-operators-q27r5\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:02 crc kubenswrapper[4944]: I1124 10:32:02.039406 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-catalog-content\") pod \"certified-operators-q27r5\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:02 crc kubenswrapper[4944]: I1124 10:32:02.039420 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-utilities\") pod \"certified-operators-q27r5\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:02 crc kubenswrapper[4944]: I1124 10:32:02.039631 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-catalog-content\") pod \"certified-operators-q27r5\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:02 crc kubenswrapper[4944]: I1124 10:32:02.083826 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgqzm\" (UniqueName: \"kubernetes.io/projected/8ffb82bd-50cd-4935-aa29-254b77054491-kube-api-access-bgqzm\") pod \"certified-operators-q27r5\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:02 crc kubenswrapper[4944]: I1124 10:32:02.095105 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4325ae-9f5e-4dba-836d-836058a44a6d","Type":"ContainerStarted","Data":"096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1"} Nov 24 10:32:02 crc kubenswrapper[4944]: I1124 10:32:02.267075 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:02 crc kubenswrapper[4944]: I1124 10:32:02.288292 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc0414f1-24b8-422c-80d8-292154d20bd4" path="/var/lib/kubelet/pods/bc0414f1-24b8-422c-80d8-292154d20bd4/volumes" Nov 24 10:32:02 crc kubenswrapper[4944]: I1124 10:32:02.805392 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q27r5"] Nov 24 10:32:03 crc kubenswrapper[4944]: I1124 10:32:03.120020 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4325ae-9f5e-4dba-836d-836058a44a6d","Type":"ContainerStarted","Data":"81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81"} Nov 24 10:32:03 crc kubenswrapper[4944]: I1124 10:32:03.123573 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q27r5" event={"ID":"8ffb82bd-50cd-4935-aa29-254b77054491","Type":"ContainerStarted","Data":"51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1"} Nov 24 10:32:03 crc kubenswrapper[4944]: I1124 10:32:03.123649 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q27r5" event={"ID":"8ffb82bd-50cd-4935-aa29-254b77054491","Type":"ContainerStarted","Data":"28927096dc0ad54d5030724f49bd0ad1e6b91fd6e7b97f98f53a82259ef16147"} Nov 24 10:32:04 crc kubenswrapper[4944]: I1124 10:32:04.141349 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4325ae-9f5e-4dba-836d-836058a44a6d","Type":"ContainerStarted","Data":"69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407"} Nov 24 10:32:04 crc kubenswrapper[4944]: I1124 10:32:04.143276 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 10:32:04 crc kubenswrapper[4944]: I1124 10:32:04.146623 4944 generic.go:334] "Generic (PLEG): container finished" podID="8ffb82bd-50cd-4935-aa29-254b77054491" containerID="51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1" exitCode=0 Nov 24 10:32:04 crc kubenswrapper[4944]: I1124 10:32:04.146665 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q27r5" event={"ID":"8ffb82bd-50cd-4935-aa29-254b77054491","Type":"ContainerDied","Data":"51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1"} Nov 24 10:32:04 crc kubenswrapper[4944]: I1124 10:32:04.146690 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q27r5" event={"ID":"8ffb82bd-50cd-4935-aa29-254b77054491","Type":"ContainerStarted","Data":"bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df"} Nov 24 10:32:04 crc kubenswrapper[4944]: I1124 10:32:04.165079 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.5178339109999999 podStartE2EDuration="5.165038203s" podCreationTimestamp="2025-11-24 10:31:59 +0000 UTC" firstStartedPulling="2025-11-24 10:32:00.057633265 +0000 UTC m=+5980.592073727" lastFinishedPulling="2025-11-24 10:32:03.704837557 +0000 UTC m=+5984.239278019" observedRunningTime="2025-11-24 10:32:04.161872613 +0000 UTC m=+5984.696313075" watchObservedRunningTime="2025-11-24 10:32:04.165038203 +0000 UTC m=+5984.699478655" Nov 24 10:32:06 crc kubenswrapper[4944]: I1124 10:32:06.193461 4944 generic.go:334] "Generic (PLEG): container finished" podID="8ffb82bd-50cd-4935-aa29-254b77054491" containerID="bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df" exitCode=0 Nov 24 10:32:06 crc kubenswrapper[4944]: I1124 10:32:06.193692 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q27r5" event={"ID":"8ffb82bd-50cd-4935-aa29-254b77054491","Type":"ContainerDied","Data":"bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df"} Nov 24 10:32:07 crc kubenswrapper[4944]: I1124 10:32:07.210415 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q27r5" event={"ID":"8ffb82bd-50cd-4935-aa29-254b77054491","Type":"ContainerStarted","Data":"ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268"} Nov 24 10:32:07 crc kubenswrapper[4944]: I1124 10:32:07.250800 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q27r5" podStartSLOduration=2.686236063 podStartE2EDuration="6.250763392s" podCreationTimestamp="2025-11-24 10:32:01 +0000 UTC" firstStartedPulling="2025-11-24 10:32:03.126369502 +0000 UTC m=+5983.660809964" lastFinishedPulling="2025-11-24 10:32:06.690896801 +0000 UTC m=+5987.225337293" observedRunningTime="2025-11-24 10:32:07.238193252 +0000 UTC m=+5987.772633744" watchObservedRunningTime="2025-11-24 10:32:07.250763392 +0000 UTC m=+5987.785203884" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.564975 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-pkg8k"] Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.566929 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pkg8k" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.577568 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-pkg8k"] Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.687335 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-2753-account-create-q5c4z"] Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.688873 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2753-account-create-q5c4z" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.696327 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.711789 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-2753-account-create-q5c4z"] Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.729463 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9vdb\" (UniqueName: \"kubernetes.io/projected/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-kube-api-access-t9vdb\") pod \"aodh-db-create-pkg8k\" (UID: \"ee0dc0ee-3410-4c15-84d4-a26d5b65405a\") " pod="openstack/aodh-db-create-pkg8k" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.730914 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-operator-scripts\") pod \"aodh-db-create-pkg8k\" (UID: \"ee0dc0ee-3410-4c15-84d4-a26d5b65405a\") " pod="openstack/aodh-db-create-pkg8k" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.832950 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-operator-scripts\") pod \"aodh-2753-account-create-q5c4z\" (UID: \"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a\") " pod="openstack/aodh-2753-account-create-q5c4z" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.833006 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dbfs\" (UniqueName: \"kubernetes.io/projected/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-kube-api-access-6dbfs\") pod \"aodh-2753-account-create-q5c4z\" (UID: \"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a\") " pod="openstack/aodh-2753-account-create-q5c4z" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.833174 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-operator-scripts\") pod \"aodh-db-create-pkg8k\" (UID: \"ee0dc0ee-3410-4c15-84d4-a26d5b65405a\") " pod="openstack/aodh-db-create-pkg8k" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.833543 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9vdb\" (UniqueName: \"kubernetes.io/projected/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-kube-api-access-t9vdb\") pod \"aodh-db-create-pkg8k\" (UID: \"ee0dc0ee-3410-4c15-84d4-a26d5b65405a\") " pod="openstack/aodh-db-create-pkg8k" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.833985 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-operator-scripts\") pod \"aodh-db-create-pkg8k\" (UID: \"ee0dc0ee-3410-4c15-84d4-a26d5b65405a\") " pod="openstack/aodh-db-create-pkg8k" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.858727 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9vdb\" (UniqueName: \"kubernetes.io/projected/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-kube-api-access-t9vdb\") pod \"aodh-db-create-pkg8k\" (UID: \"ee0dc0ee-3410-4c15-84d4-a26d5b65405a\") " pod="openstack/aodh-db-create-pkg8k" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.939926 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pkg8k" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.942101 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-operator-scripts\") pod \"aodh-2753-account-create-q5c4z\" (UID: \"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a\") " pod="openstack/aodh-2753-account-create-q5c4z" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.942183 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dbfs\" (UniqueName: \"kubernetes.io/projected/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-kube-api-access-6dbfs\") pod \"aodh-2753-account-create-q5c4z\" (UID: \"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a\") " pod="openstack/aodh-2753-account-create-q5c4z" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.945336 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-operator-scripts\") pod \"aodh-2753-account-create-q5c4z\" (UID: \"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a\") " pod="openstack/aodh-2753-account-create-q5c4z" Nov 24 10:32:10 crc kubenswrapper[4944]: I1124 10:32:10.983337 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dbfs\" (UniqueName: \"kubernetes.io/projected/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-kube-api-access-6dbfs\") pod \"aodh-2753-account-create-q5c4z\" (UID: \"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a\") " pod="openstack/aodh-2753-account-create-q5c4z" Nov 24 10:32:11 crc kubenswrapper[4944]: I1124 10:32:11.005705 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2753-account-create-q5c4z" Nov 24 10:32:11 crc kubenswrapper[4944]: I1124 10:32:11.487432 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-pkg8k"] Nov 24 10:32:11 crc kubenswrapper[4944]: I1124 10:32:11.669311 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-2753-account-create-q5c4z"] Nov 24 10:32:11 crc kubenswrapper[4944]: W1124 10:32:11.675409 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cac3c5a_0baa_4f86_ba76_6d86e8ccb41a.slice/crio-1f5395e01aedd24bdc7b1ff8c2af15d7bd7a13752fb021f052637e1c89c955fe WatchSource:0}: Error finding container 1f5395e01aedd24bdc7b1ff8c2af15d7bd7a13752fb021f052637e1c89c955fe: Status 404 returned error can't find the container with id 1f5395e01aedd24bdc7b1ff8c2af15d7bd7a13752fb021f052637e1c89c955fe Nov 24 10:32:12 crc kubenswrapper[4944]: I1124 10:32:12.044562 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-x9f7d"] Nov 24 10:32:12 crc kubenswrapper[4944]: I1124 10:32:12.055072 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-x9f7d"] Nov 24 10:32:12 crc kubenswrapper[4944]: E1124 10:32:12.187323 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cac3c5a_0baa_4f86_ba76_6d86e8ccb41a.slice/crio-5ad2a97afbfece7ee4ba94d839dd6698a890f1ee187b05601ef833f71891e1ea.scope\": RecentStats: unable to find data in memory cache]" Nov 24 10:32:12 crc kubenswrapper[4944]: I1124 10:32:12.260140 4944 generic.go:334] "Generic (PLEG): container finished" podID="ee0dc0ee-3410-4c15-84d4-a26d5b65405a" containerID="84ee30ef46f7d24d23507c81cd6777c1b4457a8f974adf970ef9128c7420b090" exitCode=0 Nov 24 10:32:12 crc kubenswrapper[4944]: I1124 10:32:12.260250 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-pkg8k" event={"ID":"ee0dc0ee-3410-4c15-84d4-a26d5b65405a","Type":"ContainerDied","Data":"84ee30ef46f7d24d23507c81cd6777c1b4457a8f974adf970ef9128c7420b090"} Nov 24 10:32:12 crc kubenswrapper[4944]: I1124 10:32:12.260416 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-pkg8k" event={"ID":"ee0dc0ee-3410-4c15-84d4-a26d5b65405a","Type":"ContainerStarted","Data":"5531a44a50975709077112b6613391ac534e28ea865e124f5af08ff658a80836"} Nov 24 10:32:12 crc kubenswrapper[4944]: I1124 10:32:12.262347 4944 generic.go:334] "Generic (PLEG): container finished" podID="8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a" containerID="5ad2a97afbfece7ee4ba94d839dd6698a890f1ee187b05601ef833f71891e1ea" exitCode=0 Nov 24 10:32:12 crc kubenswrapper[4944]: I1124 10:32:12.262390 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2753-account-create-q5c4z" event={"ID":"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a","Type":"ContainerDied","Data":"5ad2a97afbfece7ee4ba94d839dd6698a890f1ee187b05601ef833f71891e1ea"} Nov 24 10:32:12 crc kubenswrapper[4944]: I1124 10:32:12.262408 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2753-account-create-q5c4z" event={"ID":"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a","Type":"ContainerStarted","Data":"1f5395e01aedd24bdc7b1ff8c2af15d7bd7a13752fb021f052637e1c89c955fe"} Nov 24 10:32:12 crc kubenswrapper[4944]: I1124 10:32:12.268165 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:12 crc kubenswrapper[4944]: I1124 10:32:12.268314 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:12 crc kubenswrapper[4944]: I1124 10:32:12.292711 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c" path="/var/lib/kubelet/pods/dfd0d9ab-a8b9-4104-bb7b-1b1bac2bad2c/volumes" Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.033778 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-9dab-account-create-r8kk5"] Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.052141 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-mvbzx"] Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.063549 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8mh4m"] Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.073091 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-023c-account-create-9xqh8"] Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.081830 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-9dab-account-create-r8kk5"] Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.090182 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-mvbzx"] Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.098827 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8mh4m"] Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.107008 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-23ab-account-create-zv4cc"] Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.115142 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-023c-account-create-9xqh8"] Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.123384 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-23ab-account-create-zv4cc"] Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.321932 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-q27r5" podUID="8ffb82bd-50cd-4935-aa29-254b77054491" containerName="registry-server" probeResult="failure" output=< Nov 24 10:32:13 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 10:32:13 crc kubenswrapper[4944]: > Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.740645 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pkg8k" Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.747476 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2753-account-create-q5c4z" Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.816460 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-operator-scripts\") pod \"ee0dc0ee-3410-4c15-84d4-a26d5b65405a\" (UID: \"ee0dc0ee-3410-4c15-84d4-a26d5b65405a\") " Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.816526 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9vdb\" (UniqueName: \"kubernetes.io/projected/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-kube-api-access-t9vdb\") pod \"ee0dc0ee-3410-4c15-84d4-a26d5b65405a\" (UID: \"ee0dc0ee-3410-4c15-84d4-a26d5b65405a\") " Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.816893 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-operator-scripts\") pod \"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a\" (UID: \"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a\") " Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.817022 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dbfs\" (UniqueName: \"kubernetes.io/projected/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-kube-api-access-6dbfs\") pod \"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a\" (UID: \"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a\") " Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.818506 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a" (UID: "8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.818631 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee0dc0ee-3410-4c15-84d4-a26d5b65405a" (UID: "ee0dc0ee-3410-4c15-84d4-a26d5b65405a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.823454 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-kube-api-access-6dbfs" (OuterVolumeSpecName: "kube-api-access-6dbfs") pod "8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a" (UID: "8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a"). InnerVolumeSpecName "kube-api-access-6dbfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.823660 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-kube-api-access-t9vdb" (OuterVolumeSpecName: "kube-api-access-t9vdb") pod "ee0dc0ee-3410-4c15-84d4-a26d5b65405a" (UID: "ee0dc0ee-3410-4c15-84d4-a26d5b65405a"). InnerVolumeSpecName "kube-api-access-t9vdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.919672 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.919713 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dbfs\" (UniqueName: \"kubernetes.io/projected/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a-kube-api-access-6dbfs\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.919730 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:13 crc kubenswrapper[4944]: I1124 10:32:13.919743 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9vdb\" (UniqueName: \"kubernetes.io/projected/ee0dc0ee-3410-4c15-84d4-a26d5b65405a-kube-api-access-t9vdb\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:14 crc kubenswrapper[4944]: I1124 10:32:14.281684 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pkg8k" Nov 24 10:32:14 crc kubenswrapper[4944]: I1124 10:32:14.288122 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2753-account-create-q5c4z" Nov 24 10:32:14 crc kubenswrapper[4944]: I1124 10:32:14.291539 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24c9aee6-8750-4862-8030-cc249e7464b2" path="/var/lib/kubelet/pods/24c9aee6-8750-4862-8030-cc249e7464b2/volumes" Nov 24 10:32:14 crc kubenswrapper[4944]: I1124 10:32:14.292769 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4490175c-09f1-4ba1-8842-5a9a576ffce1" path="/var/lib/kubelet/pods/4490175c-09f1-4ba1-8842-5a9a576ffce1/volumes" Nov 24 10:32:14 crc kubenswrapper[4944]: I1124 10:32:14.294095 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bf65b97-ae89-462d-9bde-d694c17f8c46" path="/var/lib/kubelet/pods/6bf65b97-ae89-462d-9bde-d694c17f8c46/volumes" Nov 24 10:32:14 crc kubenswrapper[4944]: I1124 10:32:14.294633 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8269344b-cb43-4154-8550-84147c6deb7d" path="/var/lib/kubelet/pods/8269344b-cb43-4154-8550-84147c6deb7d/volumes" Nov 24 10:32:14 crc kubenswrapper[4944]: I1124 10:32:14.296039 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae14ddf-150b-4d93-9632-b6f3dff92eea" path="/var/lib/kubelet/pods/aae14ddf-150b-4d93-9632-b6f3dff92eea/volumes" Nov 24 10:32:14 crc kubenswrapper[4944]: I1124 10:32:14.297433 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-pkg8k" event={"ID":"ee0dc0ee-3410-4c15-84d4-a26d5b65405a","Type":"ContainerDied","Data":"5531a44a50975709077112b6613391ac534e28ea865e124f5af08ff658a80836"} Nov 24 10:32:14 crc kubenswrapper[4944]: I1124 10:32:14.297467 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5531a44a50975709077112b6613391ac534e28ea865e124f5af08ff658a80836" Nov 24 10:32:14 crc kubenswrapper[4944]: I1124 10:32:14.297479 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2753-account-create-q5c4z" event={"ID":"8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a","Type":"ContainerDied","Data":"1f5395e01aedd24bdc7b1ff8c2af15d7bd7a13752fb021f052637e1c89c955fe"} Nov 24 10:32:14 crc kubenswrapper[4944]: I1124 10:32:14.297491 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f5395e01aedd24bdc7b1ff8c2af15d7bd7a13752fb021f052637e1c89c955fe" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.077149 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-jr84l"] Nov 24 10:32:16 crc kubenswrapper[4944]: E1124 10:32:16.077898 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a" containerName="mariadb-account-create" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.077910 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a" containerName="mariadb-account-create" Nov 24 10:32:16 crc kubenswrapper[4944]: E1124 10:32:16.077934 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee0dc0ee-3410-4c15-84d4-a26d5b65405a" containerName="mariadb-database-create" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.077942 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0dc0ee-3410-4c15-84d4-a26d5b65405a" containerName="mariadb-database-create" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.078234 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee0dc0ee-3410-4c15-84d4-a26d5b65405a" containerName="mariadb-database-create" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.078271 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a" containerName="mariadb-account-create" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.078987 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.095600 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-jr84l"] Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.097529 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.098065 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.098187 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-vfdd8" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.098495 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.164244 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pw7z\" (UniqueName: \"kubernetes.io/projected/eda0f187-a1ab-41f0-a90b-5123d728621e-kube-api-access-5pw7z\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.164295 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-combined-ca-bundle\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.164327 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-scripts\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.164362 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-config-data\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.266016 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pw7z\" (UniqueName: \"kubernetes.io/projected/eda0f187-a1ab-41f0-a90b-5123d728621e-kube-api-access-5pw7z\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.266110 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-combined-ca-bundle\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.266155 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-scripts\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.266205 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-config-data\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.275563 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-scripts\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.277288 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-config-data\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.280929 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-combined-ca-bundle\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.281227 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pw7z\" (UniqueName: \"kubernetes.io/projected/eda0f187-a1ab-41f0-a90b-5123d728621e-kube-api-access-5pw7z\") pod \"aodh-db-sync-jr84l\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.396976 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:16 crc kubenswrapper[4944]: I1124 10:32:16.891285 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-jr84l"] Nov 24 10:32:16 crc kubenswrapper[4944]: W1124 10:32:16.892311 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeda0f187_a1ab_41f0_a90b_5123d728621e.slice/crio-fdfd4b716192abb2bdb4ca5a1cc01ddd7210d0fe017f2493b5f533a0d6c2e1cd WatchSource:0}: Error finding container fdfd4b716192abb2bdb4ca5a1cc01ddd7210d0fe017f2493b5f533a0d6c2e1cd: Status 404 returned error can't find the container with id fdfd4b716192abb2bdb4ca5a1cc01ddd7210d0fe017f2493b5f533a0d6c2e1cd Nov 24 10:32:17 crc kubenswrapper[4944]: I1124 10:32:17.320112 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jr84l" event={"ID":"eda0f187-a1ab-41f0-a90b-5123d728621e","Type":"ContainerStarted","Data":"fdfd4b716192abb2bdb4ca5a1cc01ddd7210d0fe017f2493b5f533a0d6c2e1cd"} Nov 24 10:32:20 crc kubenswrapper[4944]: I1124 10:32:20.674065 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 10:32:21 crc kubenswrapper[4944]: I1124 10:32:21.358804 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jr84l" event={"ID":"eda0f187-a1ab-41f0-a90b-5123d728621e","Type":"ContainerStarted","Data":"1bdefd3990b35b37499638c60ea1d6197ce93dc301745500112073b1189459fe"} Nov 24 10:32:21 crc kubenswrapper[4944]: I1124 10:32:21.385016 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-jr84l" podStartSLOduration=1.607775729 podStartE2EDuration="5.385000491s" podCreationTimestamp="2025-11-24 10:32:16 +0000 UTC" firstStartedPulling="2025-11-24 10:32:16.894312406 +0000 UTC m=+5997.428752868" lastFinishedPulling="2025-11-24 10:32:20.671537168 +0000 UTC m=+6001.205977630" observedRunningTime="2025-11-24 10:32:21.374479196 +0000 UTC m=+6001.908919658" watchObservedRunningTime="2025-11-24 10:32:21.385000491 +0000 UTC m=+6001.919440953" Nov 24 10:32:22 crc kubenswrapper[4944]: I1124 10:32:22.036355 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l44vr"] Nov 24 10:32:22 crc kubenswrapper[4944]: I1124 10:32:22.047655 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l44vr"] Nov 24 10:32:22 crc kubenswrapper[4944]: I1124 10:32:22.287090 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb15418b-0840-4bdb-92c0-3a8076f9fa71" path="/var/lib/kubelet/pods/bb15418b-0840-4bdb-92c0-3a8076f9fa71/volumes" Nov 24 10:32:22 crc kubenswrapper[4944]: I1124 10:32:22.313934 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:22 crc kubenswrapper[4944]: I1124 10:32:22.366464 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:22 crc kubenswrapper[4944]: I1124 10:32:22.550311 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q27r5"] Nov 24 10:32:23 crc kubenswrapper[4944]: I1124 10:32:23.378547 4944 generic.go:334] "Generic (PLEG): container finished" podID="eda0f187-a1ab-41f0-a90b-5123d728621e" containerID="1bdefd3990b35b37499638c60ea1d6197ce93dc301745500112073b1189459fe" exitCode=0 Nov 24 10:32:23 crc kubenswrapper[4944]: I1124 10:32:23.378632 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jr84l" event={"ID":"eda0f187-a1ab-41f0-a90b-5123d728621e","Type":"ContainerDied","Data":"1bdefd3990b35b37499638c60ea1d6197ce93dc301745500112073b1189459fe"} Nov 24 10:32:23 crc kubenswrapper[4944]: I1124 10:32:23.379300 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q27r5" podUID="8ffb82bd-50cd-4935-aa29-254b77054491" containerName="registry-server" containerID="cri-o://ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268" gracePeriod=2 Nov 24 10:32:23 crc kubenswrapper[4944]: I1124 10:32:23.873846 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:23 crc kubenswrapper[4944]: I1124 10:32:23.916231 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-utilities\") pod \"8ffb82bd-50cd-4935-aa29-254b77054491\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " Nov 24 10:32:23 crc kubenswrapper[4944]: I1124 10:32:23.916457 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgqzm\" (UniqueName: \"kubernetes.io/projected/8ffb82bd-50cd-4935-aa29-254b77054491-kube-api-access-bgqzm\") pod \"8ffb82bd-50cd-4935-aa29-254b77054491\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " Nov 24 10:32:23 crc kubenswrapper[4944]: I1124 10:32:23.916579 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-catalog-content\") pod \"8ffb82bd-50cd-4935-aa29-254b77054491\" (UID: \"8ffb82bd-50cd-4935-aa29-254b77054491\") " Nov 24 10:32:23 crc kubenswrapper[4944]: I1124 10:32:23.917853 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-utilities" (OuterVolumeSpecName: "utilities") pod "8ffb82bd-50cd-4935-aa29-254b77054491" (UID: "8ffb82bd-50cd-4935-aa29-254b77054491"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:32:23 crc kubenswrapper[4944]: I1124 10:32:23.923516 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ffb82bd-50cd-4935-aa29-254b77054491-kube-api-access-bgqzm" (OuterVolumeSpecName: "kube-api-access-bgqzm") pod "8ffb82bd-50cd-4935-aa29-254b77054491" (UID: "8ffb82bd-50cd-4935-aa29-254b77054491"). InnerVolumeSpecName "kube-api-access-bgqzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:32:23 crc kubenswrapper[4944]: I1124 10:32:23.973604 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ffb82bd-50cd-4935-aa29-254b77054491" (UID: "8ffb82bd-50cd-4935-aa29-254b77054491"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.019027 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.019213 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgqzm\" (UniqueName: \"kubernetes.io/projected/8ffb82bd-50cd-4935-aa29-254b77054491-kube-api-access-bgqzm\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.019295 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ffb82bd-50cd-4935-aa29-254b77054491-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.390865 4944 generic.go:334] "Generic (PLEG): container finished" podID="8ffb82bd-50cd-4935-aa29-254b77054491" containerID="ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268" exitCode=0 Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.391134 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q27r5" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.391925 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q27r5" event={"ID":"8ffb82bd-50cd-4935-aa29-254b77054491","Type":"ContainerDied","Data":"ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268"} Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.391951 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q27r5" event={"ID":"8ffb82bd-50cd-4935-aa29-254b77054491","Type":"ContainerDied","Data":"28927096dc0ad54d5030724f49bd0ad1e6b91fd6e7b97f98f53a82259ef16147"} Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.391968 4944 scope.go:117] "RemoveContainer" containerID="ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.418071 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q27r5"] Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.432590 4944 scope.go:117] "RemoveContainer" containerID="bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.436355 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q27r5"] Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.472181 4944 scope.go:117] "RemoveContainer" containerID="51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.514284 4944 scope.go:117] "RemoveContainer" containerID="ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268" Nov 24 10:32:24 crc kubenswrapper[4944]: E1124 10:32:24.514834 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268\": container with ID starting with ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268 not found: ID does not exist" containerID="ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.514886 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268"} err="failed to get container status \"ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268\": rpc error: code = NotFound desc = could not find container \"ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268\": container with ID starting with ac6d8da9b7fe339e682cedee0d5d66cd3cb24358596f55b6bcb94351aaa98268 not found: ID does not exist" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.514922 4944 scope.go:117] "RemoveContainer" containerID="bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df" Nov 24 10:32:24 crc kubenswrapper[4944]: E1124 10:32:24.515237 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df\": container with ID starting with bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df not found: ID does not exist" containerID="bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.515259 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df"} err="failed to get container status \"bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df\": rpc error: code = NotFound desc = could not find container \"bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df\": container with ID starting with bd38cfae4c0209068415d4284e4dd340913a11e3351d5eb64ca93e73e5b246df not found: ID does not exist" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.515273 4944 scope.go:117] "RemoveContainer" containerID="51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1" Nov 24 10:32:24 crc kubenswrapper[4944]: E1124 10:32:24.515453 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1\": container with ID starting with 51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1 not found: ID does not exist" containerID="51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.515476 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1"} err="failed to get container status \"51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1\": rpc error: code = NotFound desc = could not find container \"51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1\": container with ID starting with 51ba595e5e3679dd3ed69e5c8fc77de86a28018f405a8d35754782a4259614b1 not found: ID does not exist" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.762891 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.836861 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-config-data\") pod \"eda0f187-a1ab-41f0-a90b-5123d728621e\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.836991 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-combined-ca-bundle\") pod \"eda0f187-a1ab-41f0-a90b-5123d728621e\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.837073 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pw7z\" (UniqueName: \"kubernetes.io/projected/eda0f187-a1ab-41f0-a90b-5123d728621e-kube-api-access-5pw7z\") pod \"eda0f187-a1ab-41f0-a90b-5123d728621e\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.837162 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-scripts\") pod \"eda0f187-a1ab-41f0-a90b-5123d728621e\" (UID: \"eda0f187-a1ab-41f0-a90b-5123d728621e\") " Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.842295 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-scripts" (OuterVolumeSpecName: "scripts") pod "eda0f187-a1ab-41f0-a90b-5123d728621e" (UID: "eda0f187-a1ab-41f0-a90b-5123d728621e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.843809 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eda0f187-a1ab-41f0-a90b-5123d728621e-kube-api-access-5pw7z" (OuterVolumeSpecName: "kube-api-access-5pw7z") pod "eda0f187-a1ab-41f0-a90b-5123d728621e" (UID: "eda0f187-a1ab-41f0-a90b-5123d728621e"). InnerVolumeSpecName "kube-api-access-5pw7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.866591 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-config-data" (OuterVolumeSpecName: "config-data") pod "eda0f187-a1ab-41f0-a90b-5123d728621e" (UID: "eda0f187-a1ab-41f0-a90b-5123d728621e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.879933 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eda0f187-a1ab-41f0-a90b-5123d728621e" (UID: "eda0f187-a1ab-41f0-a90b-5123d728621e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.941536 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.941571 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pw7z\" (UniqueName: \"kubernetes.io/projected/eda0f187-a1ab-41f0-a90b-5123d728621e-kube-api-access-5pw7z\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.941584 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:24 crc kubenswrapper[4944]: I1124 10:32:24.941592 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda0f187-a1ab-41f0-a90b-5123d728621e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.402600 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jr84l" event={"ID":"eda0f187-a1ab-41f0-a90b-5123d728621e","Type":"ContainerDied","Data":"fdfd4b716192abb2bdb4ca5a1cc01ddd7210d0fe017f2493b5f533a0d6c2e1cd"} Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.402640 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdfd4b716192abb2bdb4ca5a1cc01ddd7210d0fe017f2493b5f533a0d6c2e1cd" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.402660 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jr84l" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.540451 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 24 10:32:25 crc kubenswrapper[4944]: E1124 10:32:25.541171 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ffb82bd-50cd-4935-aa29-254b77054491" containerName="registry-server" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.541188 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ffb82bd-50cd-4935-aa29-254b77054491" containerName="registry-server" Nov 24 10:32:25 crc kubenswrapper[4944]: E1124 10:32:25.541208 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eda0f187-a1ab-41f0-a90b-5123d728621e" containerName="aodh-db-sync" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.541214 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="eda0f187-a1ab-41f0-a90b-5123d728621e" containerName="aodh-db-sync" Nov 24 10:32:25 crc kubenswrapper[4944]: E1124 10:32:25.541230 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ffb82bd-50cd-4935-aa29-254b77054491" containerName="extract-content" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.541236 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ffb82bd-50cd-4935-aa29-254b77054491" containerName="extract-content" Nov 24 10:32:25 crc kubenswrapper[4944]: E1124 10:32:25.541255 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ffb82bd-50cd-4935-aa29-254b77054491" containerName="extract-utilities" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.541264 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ffb82bd-50cd-4935-aa29-254b77054491" containerName="extract-utilities" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.541485 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="eda0f187-a1ab-41f0-a90b-5123d728621e" containerName="aodh-db-sync" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.541512 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ffb82bd-50cd-4935-aa29-254b77054491" containerName="registry-server" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.543431 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.545777 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-vfdd8" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.547924 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.548076 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.559111 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.655285 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.655357 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvdvx\" (UniqueName: \"kubernetes.io/projected/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-kube-api-access-kvdvx\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.655639 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-config-data\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.655788 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-scripts\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.757848 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-scripts\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.758013 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.758071 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvdvx\" (UniqueName: \"kubernetes.io/projected/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-kube-api-access-kvdvx\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.758120 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-config-data\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.762524 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-scripts\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.763098 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-config-data\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.772557 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.775894 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvdvx\" (UniqueName: \"kubernetes.io/projected/8e9d68ad-ee1d-49f3-a387-62129d8b3a26-kube-api-access-kvdvx\") pod \"aodh-0\" (UID: \"8e9d68ad-ee1d-49f3-a387-62129d8b3a26\") " pod="openstack/aodh-0" Nov 24 10:32:25 crc kubenswrapper[4944]: I1124 10:32:25.904350 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 24 10:32:26 crc kubenswrapper[4944]: I1124 10:32:26.288420 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ffb82bd-50cd-4935-aa29-254b77054491" path="/var/lib/kubelet/pods/8ffb82bd-50cd-4935-aa29-254b77054491/volumes" Nov 24 10:32:26 crc kubenswrapper[4944]: I1124 10:32:26.405508 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 24 10:32:26 crc kubenswrapper[4944]: W1124 10:32:26.415436 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e9d68ad_ee1d_49f3_a387_62129d8b3a26.slice/crio-77134d77834bb75fd679a363b2c3547faafd7c04d13fc0e4750cfc3f5853ee83 WatchSource:0}: Error finding container 77134d77834bb75fd679a363b2c3547faafd7c04d13fc0e4750cfc3f5853ee83: Status 404 returned error can't find the container with id 77134d77834bb75fd679a363b2c3547faafd7c04d13fc0e4750cfc3f5853ee83 Nov 24 10:32:27 crc kubenswrapper[4944]: I1124 10:32:27.430224 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8e9d68ad-ee1d-49f3-a387-62129d8b3a26","Type":"ContainerStarted","Data":"28939244bd9e01fb4b0ef8e613de6105d2a2c6130322af4619b93928052c58ad"} Nov 24 10:32:27 crc kubenswrapper[4944]: I1124 10:32:27.430554 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8e9d68ad-ee1d-49f3-a387-62129d8b3a26","Type":"ContainerStarted","Data":"77134d77834bb75fd679a363b2c3547faafd7c04d13fc0e4750cfc3f5853ee83"} Nov 24 10:32:27 crc kubenswrapper[4944]: I1124 10:32:27.649706 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:32:27 crc kubenswrapper[4944]: I1124 10:32:27.650013 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="ceilometer-central-agent" containerID="cri-o://543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9" gracePeriod=30 Nov 24 10:32:27 crc kubenswrapper[4944]: I1124 10:32:27.650135 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="sg-core" containerID="cri-o://81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81" gracePeriod=30 Nov 24 10:32:27 crc kubenswrapper[4944]: I1124 10:32:27.650087 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="proxy-httpd" containerID="cri-o://69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407" gracePeriod=30 Nov 24 10:32:27 crc kubenswrapper[4944]: I1124 10:32:27.650186 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="ceilometer-notification-agent" containerID="cri-o://096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1" gracePeriod=30 Nov 24 10:32:27 crc kubenswrapper[4944]: I1124 10:32:27.656363 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.1.131:3000/\": EOF" Nov 24 10:32:28 crc kubenswrapper[4944]: I1124 10:32:28.460735 4944 generic.go:334] "Generic (PLEG): container finished" podID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerID="69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407" exitCode=0 Nov 24 10:32:28 crc kubenswrapper[4944]: I1124 10:32:28.461313 4944 generic.go:334] "Generic (PLEG): container finished" podID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerID="81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81" exitCode=2 Nov 24 10:32:28 crc kubenswrapper[4944]: I1124 10:32:28.461331 4944 generic.go:334] "Generic (PLEG): container finished" podID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerID="543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9" exitCode=0 Nov 24 10:32:28 crc kubenswrapper[4944]: I1124 10:32:28.461365 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4325ae-9f5e-4dba-836d-836058a44a6d","Type":"ContainerDied","Data":"69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407"} Nov 24 10:32:28 crc kubenswrapper[4944]: I1124 10:32:28.461403 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4325ae-9f5e-4dba-836d-836058a44a6d","Type":"ContainerDied","Data":"81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81"} Nov 24 10:32:28 crc kubenswrapper[4944]: I1124 10:32:28.461424 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4325ae-9f5e-4dba-836d-836058a44a6d","Type":"ContainerDied","Data":"543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9"} Nov 24 10:32:29 crc kubenswrapper[4944]: I1124 10:32:29.471920 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8e9d68ad-ee1d-49f3-a387-62129d8b3a26","Type":"ContainerStarted","Data":"bb5cbae663462e69b895df5cb6dfd7fd88e959e0fd7054f5e7f176ca2e04ad6e"} Nov 24 10:32:29 crc kubenswrapper[4944]: I1124 10:32:29.482696 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.1.131:3000/\": dial tcp 10.217.1.131:3000: connect: connection refused" Nov 24 10:32:30 crc kubenswrapper[4944]: I1124 10:32:30.485397 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8e9d68ad-ee1d-49f3-a387-62129d8b3a26","Type":"ContainerStarted","Data":"d3aa5a4ce0594da695efb208796b1aa444fdecf1f011d18ab49abad5e1719d73"} Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.480713 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fcjpj"] Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.484380 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.505339 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fcjpj"] Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.511260 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8e9d68ad-ee1d-49f3-a387-62129d8b3a26","Type":"ContainerStarted","Data":"b623d4ee6b640e7fa63ea2efd95140c727ffb2e8b52dcdeb33b916d7782ff0c2"} Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.544000 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.113853623 podStartE2EDuration="6.54396716s" podCreationTimestamp="2025-11-24 10:32:25 +0000 UTC" firstStartedPulling="2025-11-24 10:32:26.424460476 +0000 UTC m=+6006.958900948" lastFinishedPulling="2025-11-24 10:32:30.854574023 +0000 UTC m=+6011.389014485" observedRunningTime="2025-11-24 10:32:31.53326738 +0000 UTC m=+6012.067707842" watchObservedRunningTime="2025-11-24 10:32:31.54396716 +0000 UTC m=+6012.078407622" Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.584324 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-utilities\") pod \"redhat-marketplace-fcjpj\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.584489 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t58f7\" (UniqueName: \"kubernetes.io/projected/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-kube-api-access-t58f7\") pod \"redhat-marketplace-fcjpj\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.584585 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-catalog-content\") pod \"redhat-marketplace-fcjpj\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.686247 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-utilities\") pod \"redhat-marketplace-fcjpj\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.686378 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t58f7\" (UniqueName: \"kubernetes.io/projected/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-kube-api-access-t58f7\") pod \"redhat-marketplace-fcjpj\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.686477 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-catalog-content\") pod \"redhat-marketplace-fcjpj\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.686701 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-utilities\") pod \"redhat-marketplace-fcjpj\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.686924 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-catalog-content\") pod \"redhat-marketplace-fcjpj\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.703336 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t58f7\" (UniqueName: \"kubernetes.io/projected/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-kube-api-access-t58f7\") pod \"redhat-marketplace-fcjpj\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:31 crc kubenswrapper[4944]: I1124 10:32:31.807026 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:32 crc kubenswrapper[4944]: I1124 10:32:32.273664 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fcjpj"] Nov 24 10:32:32 crc kubenswrapper[4944]: W1124 10:32:32.276614 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3930c1d8_0a2e_4035_8ded_b6dd2d2a0b02.slice/crio-25ee29cc8063d07de7d5e22e6b9a83a26f4236b6c2fdd436cc0328f13b1e244e WatchSource:0}: Error finding container 25ee29cc8063d07de7d5e22e6b9a83a26f4236b6c2fdd436cc0328f13b1e244e: Status 404 returned error can't find the container with id 25ee29cc8063d07de7d5e22e6b9a83a26f4236b6c2fdd436cc0328f13b1e244e Nov 24 10:32:32 crc kubenswrapper[4944]: I1124 10:32:32.523681 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcjpj" event={"ID":"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02","Type":"ContainerStarted","Data":"d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558"} Nov 24 10:32:32 crc kubenswrapper[4944]: I1124 10:32:32.523731 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcjpj" event={"ID":"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02","Type":"ContainerStarted","Data":"25ee29cc8063d07de7d5e22e6b9a83a26f4236b6c2fdd436cc0328f13b1e244e"} Nov 24 10:32:32 crc kubenswrapper[4944]: E1124 10:32:32.718424 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3930c1d8_0a2e_4035_8ded_b6dd2d2a0b02.slice/crio-conmon-d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558.scope\": RecentStats: unable to find data in memory cache]" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.256969 4944 scope.go:117] "RemoveContainer" containerID="5d7739bd33ffb92b9bbf110c7a163cd4f76dd6925a8f1631772040af87b2bfd3" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.282358 4944 scope.go:117] "RemoveContainer" containerID="2b1490de7e5d194d5a177b2bcb42fd34098c29b5bf9de2e5a82c0b6430805910" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.359781 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.369564 4944 scope.go:117] "RemoveContainer" containerID="083d4742ebc41b7ca3957c3ef4d36ce42a9b5f560f247bb1dbc1572aba343811" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.416530 4944 scope.go:117] "RemoveContainer" containerID="78af2b150e96a2d80d31547eed468b9b38c6965eab76d0cc7e2e8cb47334d259" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.427610 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-sg-core-conf-yaml\") pod \"2d4325ae-9f5e-4dba-836d-836058a44a6d\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.427653 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-combined-ca-bundle\") pod \"2d4325ae-9f5e-4dba-836d-836058a44a6d\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.427755 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjznp\" (UniqueName: \"kubernetes.io/projected/2d4325ae-9f5e-4dba-836d-836058a44a6d-kube-api-access-vjznp\") pod \"2d4325ae-9f5e-4dba-836d-836058a44a6d\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.428528 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-scripts\") pod \"2d4325ae-9f5e-4dba-836d-836058a44a6d\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.428841 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-log-httpd\") pod \"2d4325ae-9f5e-4dba-836d-836058a44a6d\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.428874 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-config-data\") pod \"2d4325ae-9f5e-4dba-836d-836058a44a6d\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.428914 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-run-httpd\") pod \"2d4325ae-9f5e-4dba-836d-836058a44a6d\" (UID: \"2d4325ae-9f5e-4dba-836d-836058a44a6d\") " Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.429396 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2d4325ae-9f5e-4dba-836d-836058a44a6d" (UID: "2d4325ae-9f5e-4dba-836d-836058a44a6d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.429660 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2d4325ae-9f5e-4dba-836d-836058a44a6d" (UID: "2d4325ae-9f5e-4dba-836d-836058a44a6d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.433186 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-scripts" (OuterVolumeSpecName: "scripts") pod "2d4325ae-9f5e-4dba-836d-836058a44a6d" (UID: "2d4325ae-9f5e-4dba-836d-836058a44a6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.433454 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d4325ae-9f5e-4dba-836d-836058a44a6d-kube-api-access-vjznp" (OuterVolumeSpecName: "kube-api-access-vjznp") pod "2d4325ae-9f5e-4dba-836d-836058a44a6d" (UID: "2d4325ae-9f5e-4dba-836d-836058a44a6d"). InnerVolumeSpecName "kube-api-access-vjznp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.456721 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2d4325ae-9f5e-4dba-836d-836058a44a6d" (UID: "2d4325ae-9f5e-4dba-836d-836058a44a6d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.457396 4944 scope.go:117] "RemoveContainer" containerID="4269de9f24f4eb215c2e66c0fc50eeb78e34b2d967b36452fea0d2f63b6121f1" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.481480 4944 scope.go:117] "RemoveContainer" containerID="82e61cc90223698151e9f7f0d34f7b6f6d39f25a0972101eb1dc88b883f153e8" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.519061 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d4325ae-9f5e-4dba-836d-836058a44a6d" (UID: "2d4325ae-9f5e-4dba-836d-836058a44a6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.526434 4944 scope.go:117] "RemoveContainer" containerID="002f641c5545a01ed995891d667a5662ec851ca2b44e2266a4daef8073bfc478" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.534249 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjznp\" (UniqueName: \"kubernetes.io/projected/2d4325ae-9f5e-4dba-836d-836058a44a6d-kube-api-access-vjznp\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.534280 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.534290 4944 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.534299 4944 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2d4325ae-9f5e-4dba-836d-836058a44a6d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.534309 4944 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.534320 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.543565 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-config-data" (OuterVolumeSpecName: "config-data") pod "2d4325ae-9f5e-4dba-836d-836058a44a6d" (UID: "2d4325ae-9f5e-4dba-836d-836058a44a6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.543805 4944 generic.go:334] "Generic (PLEG): container finished" podID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerID="096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1" exitCode=0 Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.543868 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4325ae-9f5e-4dba-836d-836058a44a6d","Type":"ContainerDied","Data":"096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1"} Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.543932 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2d4325ae-9f5e-4dba-836d-836058a44a6d","Type":"ContainerDied","Data":"67d1b13cbd405eec579abcc5af453604eceb7a49a10bb8d78d04cc5458cecc53"} Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.543955 4944 scope.go:117] "RemoveContainer" containerID="69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.543897 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.563204 4944 generic.go:334] "Generic (PLEG): container finished" podID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" containerID="d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558" exitCode=0 Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.563407 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcjpj" event={"ID":"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02","Type":"ContainerDied","Data":"d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558"} Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.584082 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.585133 4944 scope.go:117] "RemoveContainer" containerID="81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.595586 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.625946 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:32:33 crc kubenswrapper[4944]: E1124 10:32:33.626389 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="sg-core" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.626405 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="sg-core" Nov 24 10:32:33 crc kubenswrapper[4944]: E1124 10:32:33.626441 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="ceilometer-central-agent" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.626448 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="ceilometer-central-agent" Nov 24 10:32:33 crc kubenswrapper[4944]: E1124 10:32:33.626468 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="proxy-httpd" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.626474 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="proxy-httpd" Nov 24 10:32:33 crc kubenswrapper[4944]: E1124 10:32:33.626486 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="ceilometer-notification-agent" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.626492 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="ceilometer-notification-agent" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.626674 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="ceilometer-notification-agent" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.626698 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="sg-core" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.626719 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="ceilometer-central-agent" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.626732 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" containerName="proxy-httpd" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.628593 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.634116 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.636023 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.637365 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-scripts\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.637600 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk526\" (UniqueName: \"kubernetes.io/projected/41b72775-aac7-4695-bd24-de5904aeebdb-kube-api-access-bk526\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.637801 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.638001 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-config-data\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.638194 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-run-httpd\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.638348 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-log-httpd\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.638479 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.638629 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4325ae-9f5e-4dba-836d-836058a44a6d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.666312 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.667377 4944 scope.go:117] "RemoveContainer" containerID="096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.705464 4944 scope.go:117] "RemoveContainer" containerID="543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.728387 4944 scope.go:117] "RemoveContainer" containerID="69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407" Nov 24 10:32:33 crc kubenswrapper[4944]: E1124 10:32:33.728869 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407\": container with ID starting with 69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407 not found: ID does not exist" containerID="69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.728952 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407"} err="failed to get container status \"69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407\": rpc error: code = NotFound desc = could not find container \"69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407\": container with ID starting with 69eea962c58eb9b157aa7c717f5fe412882ae8b231aaa9fbf961bc10bd3c2407 not found: ID does not exist" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.729032 4944 scope.go:117] "RemoveContainer" containerID="81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81" Nov 24 10:32:33 crc kubenswrapper[4944]: E1124 10:32:33.729438 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81\": container with ID starting with 81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81 not found: ID does not exist" containerID="81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.729513 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81"} err="failed to get container status \"81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81\": rpc error: code = NotFound desc = could not find container \"81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81\": container with ID starting with 81a91c0da0fb2586fed48e51b54df7e7f01bde5a9e718427bd15e3dc2ad75f81 not found: ID does not exist" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.729572 4944 scope.go:117] "RemoveContainer" containerID="096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1" Nov 24 10:32:33 crc kubenswrapper[4944]: E1124 10:32:33.729890 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1\": container with ID starting with 096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1 not found: ID does not exist" containerID="096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.729978 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1"} err="failed to get container status \"096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1\": rpc error: code = NotFound desc = could not find container \"096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1\": container with ID starting with 096f045f9600b3d11166da8d04379dd7772591e6976cc24e2c325ae7217fedf1 not found: ID does not exist" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.730078 4944 scope.go:117] "RemoveContainer" containerID="543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9" Nov 24 10:32:33 crc kubenswrapper[4944]: E1124 10:32:33.730514 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9\": container with ID starting with 543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9 not found: ID does not exist" containerID="543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.730538 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9"} err="failed to get container status \"543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9\": rpc error: code = NotFound desc = could not find container \"543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9\": container with ID starting with 543502e6ae5aa8116c33528536ae08ed950029f676c8ed05f16c6f7ca841fcb9 not found: ID does not exist" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.739610 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk526\" (UniqueName: \"kubernetes.io/projected/41b72775-aac7-4695-bd24-de5904aeebdb-kube-api-access-bk526\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.740184 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.740329 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-config-data\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.740459 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-run-httpd\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.740542 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-log-httpd\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.740621 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.740766 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-scripts\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.741442 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-log-httpd\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.741711 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-run-httpd\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.745218 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.745309 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-scripts\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.745835 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-config-data\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.756829 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.759390 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk526\" (UniqueName: \"kubernetes.io/projected/41b72775-aac7-4695-bd24-de5904aeebdb-kube-api-access-bk526\") pod \"ceilometer-0\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " pod="openstack/ceilometer-0" Nov 24 10:32:33 crc kubenswrapper[4944]: I1124 10:32:33.951210 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:32:34 crc kubenswrapper[4944]: I1124 10:32:34.312029 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d4325ae-9f5e-4dba-836d-836058a44a6d" path="/var/lib/kubelet/pods/2d4325ae-9f5e-4dba-836d-836058a44a6d/volumes" Nov 24 10:32:34 crc kubenswrapper[4944]: I1124 10:32:34.493568 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:32:34 crc kubenswrapper[4944]: W1124 10:32:34.505985 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41b72775_aac7_4695_bd24_de5904aeebdb.slice/crio-783fe64ac1dea47436a8a689dd5fdaab4270374369a6a6276f6ff318458188d3 WatchSource:0}: Error finding container 783fe64ac1dea47436a8a689dd5fdaab4270374369a6a6276f6ff318458188d3: Status 404 returned error can't find the container with id 783fe64ac1dea47436a8a689dd5fdaab4270374369a6a6276f6ff318458188d3 Nov 24 10:32:34 crc kubenswrapper[4944]: I1124 10:32:34.578876 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41b72775-aac7-4695-bd24-de5904aeebdb","Type":"ContainerStarted","Data":"783fe64ac1dea47436a8a689dd5fdaab4270374369a6a6276f6ff318458188d3"} Nov 24 10:32:34 crc kubenswrapper[4944]: I1124 10:32:34.581263 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcjpj" event={"ID":"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02","Type":"ContainerStarted","Data":"0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a"} Nov 24 10:32:35 crc kubenswrapper[4944]: I1124 10:32:35.591130 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41b72775-aac7-4695-bd24-de5904aeebdb","Type":"ContainerStarted","Data":"51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2"} Nov 24 10:32:35 crc kubenswrapper[4944]: I1124 10:32:35.593562 4944 generic.go:334] "Generic (PLEG): container finished" podID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" containerID="0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a" exitCode=0 Nov 24 10:32:35 crc kubenswrapper[4944]: I1124 10:32:35.593601 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcjpj" event={"ID":"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02","Type":"ContainerDied","Data":"0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a"} Nov 24 10:32:36 crc kubenswrapper[4944]: I1124 10:32:36.607374 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcjpj" event={"ID":"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02","Type":"ContainerStarted","Data":"5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb"} Nov 24 10:32:36 crc kubenswrapper[4944]: I1124 10:32:36.612018 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41b72775-aac7-4695-bd24-de5904aeebdb","Type":"ContainerStarted","Data":"0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0"} Nov 24 10:32:36 crc kubenswrapper[4944]: I1124 10:32:36.612101 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41b72775-aac7-4695-bd24-de5904aeebdb","Type":"ContainerStarted","Data":"4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993"} Nov 24 10:32:36 crc kubenswrapper[4944]: I1124 10:32:36.626876 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fcjpj" podStartSLOduration=3.242280499 podStartE2EDuration="5.626858998s" podCreationTimestamp="2025-11-24 10:32:31 +0000 UTC" firstStartedPulling="2025-11-24 10:32:33.58546351 +0000 UTC m=+6014.119903972" lastFinishedPulling="2025-11-24 10:32:35.970042009 +0000 UTC m=+6016.504482471" observedRunningTime="2025-11-24 10:32:36.62127042 +0000 UTC m=+6017.155710902" watchObservedRunningTime="2025-11-24 10:32:36.626858998 +0000 UTC m=+6017.161299460" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.362669 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-t2nc4"] Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.364426 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t2nc4" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.374463 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-t2nc4"] Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.469533 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-9d4b-account-create-8pw4q"] Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.471424 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9d4b-account-create-8pw4q" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.478414 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.493067 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-9d4b-account-create-8pw4q"] Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.527494 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzghz\" (UniqueName: \"kubernetes.io/projected/0e1d1afa-085a-4cc1-b772-9880bb17630e-kube-api-access-kzghz\") pod \"manila-db-create-t2nc4\" (UID: \"0e1d1afa-085a-4cc1-b772-9880bb17630e\") " pod="openstack/manila-db-create-t2nc4" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.527596 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e1d1afa-085a-4cc1-b772-9880bb17630e-operator-scripts\") pod \"manila-db-create-t2nc4\" (UID: \"0e1d1afa-085a-4cc1-b772-9880bb17630e\") " pod="openstack/manila-db-create-t2nc4" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.629842 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-operator-scripts\") pod \"manila-9d4b-account-create-8pw4q\" (UID: \"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a\") " pod="openstack/manila-9d4b-account-create-8pw4q" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.630077 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzghz\" (UniqueName: \"kubernetes.io/projected/0e1d1afa-085a-4cc1-b772-9880bb17630e-kube-api-access-kzghz\") pod \"manila-db-create-t2nc4\" (UID: \"0e1d1afa-085a-4cc1-b772-9880bb17630e\") " pod="openstack/manila-db-create-t2nc4" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.630146 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv2qw\" (UniqueName: \"kubernetes.io/projected/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-kube-api-access-tv2qw\") pod \"manila-9d4b-account-create-8pw4q\" (UID: \"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a\") " pod="openstack/manila-9d4b-account-create-8pw4q" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.630214 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e1d1afa-085a-4cc1-b772-9880bb17630e-operator-scripts\") pod \"manila-db-create-t2nc4\" (UID: \"0e1d1afa-085a-4cc1-b772-9880bb17630e\") " pod="openstack/manila-db-create-t2nc4" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.631101 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e1d1afa-085a-4cc1-b772-9880bb17630e-operator-scripts\") pod \"manila-db-create-t2nc4\" (UID: \"0e1d1afa-085a-4cc1-b772-9880bb17630e\") " pod="openstack/manila-db-create-t2nc4" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.656173 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzghz\" (UniqueName: \"kubernetes.io/projected/0e1d1afa-085a-4cc1-b772-9880bb17630e-kube-api-access-kzghz\") pod \"manila-db-create-t2nc4\" (UID: \"0e1d1afa-085a-4cc1-b772-9880bb17630e\") " pod="openstack/manila-db-create-t2nc4" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.685242 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t2nc4" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.732321 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv2qw\" (UniqueName: \"kubernetes.io/projected/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-kube-api-access-tv2qw\") pod \"manila-9d4b-account-create-8pw4q\" (UID: \"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a\") " pod="openstack/manila-9d4b-account-create-8pw4q" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.732449 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-operator-scripts\") pod \"manila-9d4b-account-create-8pw4q\" (UID: \"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a\") " pod="openstack/manila-9d4b-account-create-8pw4q" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.733398 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-operator-scripts\") pod \"manila-9d4b-account-create-8pw4q\" (UID: \"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a\") " pod="openstack/manila-9d4b-account-create-8pw4q" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.754625 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv2qw\" (UniqueName: \"kubernetes.io/projected/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-kube-api-access-tv2qw\") pod \"manila-9d4b-account-create-8pw4q\" (UID: \"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a\") " pod="openstack/manila-9d4b-account-create-8pw4q" Nov 24 10:32:37 crc kubenswrapper[4944]: I1124 10:32:37.790694 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9d4b-account-create-8pw4q" Nov 24 10:32:38 crc kubenswrapper[4944]: I1124 10:32:38.237186 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-t2nc4"] Nov 24 10:32:38 crc kubenswrapper[4944]: W1124 10:32:38.245671 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e1d1afa_085a_4cc1_b772_9880bb17630e.slice/crio-c1293c3fa3af328e92b67aec24effc41904cf1e6e510400a98c71e2adbd5dcc8 WatchSource:0}: Error finding container c1293c3fa3af328e92b67aec24effc41904cf1e6e510400a98c71e2adbd5dcc8: Status 404 returned error can't find the container with id c1293c3fa3af328e92b67aec24effc41904cf1e6e510400a98c71e2adbd5dcc8 Nov 24 10:32:38 crc kubenswrapper[4944]: I1124 10:32:38.510600 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-9d4b-account-create-8pw4q"] Nov 24 10:32:38 crc kubenswrapper[4944]: W1124 10:32:38.517871 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc88c41d9_7b3e_45e9_8a93_a6efeb696c1a.slice/crio-2acc9829898a5351f0b79b2f42bc3e63f59760e60d91e0982ef64c0876f4528a WatchSource:0}: Error finding container 2acc9829898a5351f0b79b2f42bc3e63f59760e60d91e0982ef64c0876f4528a: Status 404 returned error can't find the container with id 2acc9829898a5351f0b79b2f42bc3e63f59760e60d91e0982ef64c0876f4528a Nov 24 10:32:38 crc kubenswrapper[4944]: I1124 10:32:38.637851 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-t2nc4" event={"ID":"0e1d1afa-085a-4cc1-b772-9880bb17630e","Type":"ContainerStarted","Data":"5a38ae420edb5d72efdd021fb55317d671c8b93642d008d2a5b43b45bca5bfb7"} Nov 24 10:32:38 crc kubenswrapper[4944]: I1124 10:32:38.637896 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-t2nc4" event={"ID":"0e1d1afa-085a-4cc1-b772-9880bb17630e","Type":"ContainerStarted","Data":"c1293c3fa3af328e92b67aec24effc41904cf1e6e510400a98c71e2adbd5dcc8"} Nov 24 10:32:38 crc kubenswrapper[4944]: I1124 10:32:38.643282 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9d4b-account-create-8pw4q" event={"ID":"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a","Type":"ContainerStarted","Data":"2acc9829898a5351f0b79b2f42bc3e63f59760e60d91e0982ef64c0876f4528a"} Nov 24 10:32:38 crc kubenswrapper[4944]: I1124 10:32:38.664217 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-create-t2nc4" podStartSLOduration=1.664187165 podStartE2EDuration="1.664187165s" podCreationTimestamp="2025-11-24 10:32:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:32:38.652443012 +0000 UTC m=+6019.186883474" watchObservedRunningTime="2025-11-24 10:32:38.664187165 +0000 UTC m=+6019.198627627" Nov 24 10:32:39 crc kubenswrapper[4944]: I1124 10:32:39.655574 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41b72775-aac7-4695-bd24-de5904aeebdb","Type":"ContainerStarted","Data":"849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d"} Nov 24 10:32:39 crc kubenswrapper[4944]: I1124 10:32:39.656001 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 10:32:39 crc kubenswrapper[4944]: I1124 10:32:39.657577 4944 generic.go:334] "Generic (PLEG): container finished" podID="c88c41d9-7b3e-45e9-8a93-a6efeb696c1a" containerID="4fafff5ddb63d7ff80bc2a407fe14081b6cf4503524602e3393ea68dc9b2ab9d" exitCode=0 Nov 24 10:32:39 crc kubenswrapper[4944]: I1124 10:32:39.657750 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9d4b-account-create-8pw4q" event={"ID":"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a","Type":"ContainerDied","Data":"4fafff5ddb63d7ff80bc2a407fe14081b6cf4503524602e3393ea68dc9b2ab9d"} Nov 24 10:32:39 crc kubenswrapper[4944]: I1124 10:32:39.668678 4944 generic.go:334] "Generic (PLEG): container finished" podID="0e1d1afa-085a-4cc1-b772-9880bb17630e" containerID="5a38ae420edb5d72efdd021fb55317d671c8b93642d008d2a5b43b45bca5bfb7" exitCode=0 Nov 24 10:32:39 crc kubenswrapper[4944]: I1124 10:32:39.668741 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-t2nc4" event={"ID":"0e1d1afa-085a-4cc1-b772-9880bb17630e","Type":"ContainerDied","Data":"5a38ae420edb5d72efdd021fb55317d671c8b93642d008d2a5b43b45bca5bfb7"} Nov 24 10:32:39 crc kubenswrapper[4944]: I1124 10:32:39.677424 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.716592156 podStartE2EDuration="6.677404556s" podCreationTimestamp="2025-11-24 10:32:33 +0000 UTC" firstStartedPulling="2025-11-24 10:32:34.508593542 +0000 UTC m=+6015.043034004" lastFinishedPulling="2025-11-24 10:32:38.469405942 +0000 UTC m=+6019.003846404" observedRunningTime="2025-11-24 10:32:39.675174695 +0000 UTC m=+6020.209615157" watchObservedRunningTime="2025-11-24 10:32:39.677404556 +0000 UTC m=+6020.211845018" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.047098 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fm68v"] Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.058252 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fm68v"] Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.150887 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9d4b-account-create-8pw4q" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.157695 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t2nc4" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.318867 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv2qw\" (UniqueName: \"kubernetes.io/projected/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-kube-api-access-tv2qw\") pod \"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a\" (UID: \"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a\") " Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.319619 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e1d1afa-085a-4cc1-b772-9880bb17630e-operator-scripts\") pod \"0e1d1afa-085a-4cc1-b772-9880bb17630e\" (UID: \"0e1d1afa-085a-4cc1-b772-9880bb17630e\") " Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.319666 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzghz\" (UniqueName: \"kubernetes.io/projected/0e1d1afa-085a-4cc1-b772-9880bb17630e-kube-api-access-kzghz\") pod \"0e1d1afa-085a-4cc1-b772-9880bb17630e\" (UID: \"0e1d1afa-085a-4cc1-b772-9880bb17630e\") " Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.319708 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-operator-scripts\") pod \"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a\" (UID: \"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a\") " Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.319965 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e1d1afa-085a-4cc1-b772-9880bb17630e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e1d1afa-085a-4cc1-b772-9880bb17630e" (UID: "0e1d1afa-085a-4cc1-b772-9880bb17630e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.320127 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e1d1afa-085a-4cc1-b772-9880bb17630e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.320552 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c88c41d9-7b3e-45e9-8a93-a6efeb696c1a" (UID: "c88c41d9-7b3e-45e9-8a93-a6efeb696c1a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.325452 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e1d1afa-085a-4cc1-b772-9880bb17630e-kube-api-access-kzghz" (OuterVolumeSpecName: "kube-api-access-kzghz") pod "0e1d1afa-085a-4cc1-b772-9880bb17630e" (UID: "0e1d1afa-085a-4cc1-b772-9880bb17630e"). InnerVolumeSpecName "kube-api-access-kzghz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.326306 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-kube-api-access-tv2qw" (OuterVolumeSpecName: "kube-api-access-tv2qw") pod "c88c41d9-7b3e-45e9-8a93-a6efeb696c1a" (UID: "c88c41d9-7b3e-45e9-8a93-a6efeb696c1a"). InnerVolumeSpecName "kube-api-access-tv2qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.421664 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv2qw\" (UniqueName: \"kubernetes.io/projected/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-kube-api-access-tv2qw\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.421694 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzghz\" (UniqueName: \"kubernetes.io/projected/0e1d1afa-085a-4cc1-b772-9880bb17630e-kube-api-access-kzghz\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.421703 4944 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.695498 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-t2nc4" event={"ID":"0e1d1afa-085a-4cc1-b772-9880bb17630e","Type":"ContainerDied","Data":"c1293c3fa3af328e92b67aec24effc41904cf1e6e510400a98c71e2adbd5dcc8"} Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.695543 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1293c3fa3af328e92b67aec24effc41904cf1e6e510400a98c71e2adbd5dcc8" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.696092 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-t2nc4" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.698442 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9d4b-account-create-8pw4q" event={"ID":"c88c41d9-7b3e-45e9-8a93-a6efeb696c1a","Type":"ContainerDied","Data":"2acc9829898a5351f0b79b2f42bc3e63f59760e60d91e0982ef64c0876f4528a"} Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.698555 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2acc9829898a5351f0b79b2f42bc3e63f59760e60d91e0982ef64c0876f4528a" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.698465 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9d4b-account-create-8pw4q" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.807447 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.807791 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:41 crc kubenswrapper[4944]: I1124 10:32:41.857256 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.028718 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-bb2gt"] Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.039632 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-bb2gt"] Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.287644 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73011835-f092-4f8c-8ec2-d5928bb99c97" path="/var/lib/kubelet/pods/73011835-f092-4f8c-8ec2-d5928bb99c97/volumes" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.290421 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1bbbe09-dbf8-434a-8615-849e02e7873a" path="/var/lib/kubelet/pods/f1bbbe09-dbf8-434a-8615-849e02e7873a/volumes" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.769399 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.829310 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-9b42b"] Nov 24 10:32:42 crc kubenswrapper[4944]: E1124 10:32:42.829836 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e1d1afa-085a-4cc1-b772-9880bb17630e" containerName="mariadb-database-create" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.829855 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e1d1afa-085a-4cc1-b772-9880bb17630e" containerName="mariadb-database-create" Nov 24 10:32:42 crc kubenswrapper[4944]: E1124 10:32:42.829909 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c88c41d9-7b3e-45e9-8a93-a6efeb696c1a" containerName="mariadb-account-create" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.829918 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c88c41d9-7b3e-45e9-8a93-a6efeb696c1a" containerName="mariadb-account-create" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.830160 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="c88c41d9-7b3e-45e9-8a93-a6efeb696c1a" containerName="mariadb-account-create" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.830194 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e1d1afa-085a-4cc1-b772-9880bb17630e" containerName="mariadb-database-create" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.831116 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.832838 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.834676 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-94x8g" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.843241 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-9b42b"] Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.951348 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-config-data\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.951409 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hl66\" (UniqueName: \"kubernetes.io/projected/761360e0-d8c9-4a45-96a3-c90e19f721fc-kube-api-access-7hl66\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.951441 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-job-config-data\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:42 crc kubenswrapper[4944]: I1124 10:32:42.951488 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-combined-ca-bundle\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.053690 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-config-data\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.053739 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hl66\" (UniqueName: \"kubernetes.io/projected/761360e0-d8c9-4a45-96a3-c90e19f721fc-kube-api-access-7hl66\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.053761 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-job-config-data\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.053800 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-combined-ca-bundle\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.058887 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-job-config-data\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.061824 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-config-data\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.062361 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-combined-ca-bundle\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.069315 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hl66\" (UniqueName: \"kubernetes.io/projected/761360e0-d8c9-4a45-96a3-c90e19f721fc-kube-api-access-7hl66\") pod \"manila-db-sync-9b42b\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.155198 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.294619 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fcjpj"] Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.763123 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-9b42b"] Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.901432 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8zg45"] Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.907740 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:43 crc kubenswrapper[4944]: I1124 10:32:43.932333 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8zg45"] Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.075351 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-utilities\") pod \"community-operators-8zg45\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.075459 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwq9r\" (UniqueName: \"kubernetes.io/projected/18a76545-dbb6-4e83-ad53-de6924dc985e-kube-api-access-fwq9r\") pod \"community-operators-8zg45\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.075501 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-catalog-content\") pod \"community-operators-8zg45\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.177902 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-utilities\") pod \"community-operators-8zg45\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.177981 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwq9r\" (UniqueName: \"kubernetes.io/projected/18a76545-dbb6-4e83-ad53-de6924dc985e-kube-api-access-fwq9r\") pod \"community-operators-8zg45\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.178039 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-catalog-content\") pod \"community-operators-8zg45\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.178783 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-catalog-content\") pod \"community-operators-8zg45\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.179484 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-utilities\") pod \"community-operators-8zg45\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.203668 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwq9r\" (UniqueName: \"kubernetes.io/projected/18a76545-dbb6-4e83-ad53-de6924dc985e-kube-api-access-fwq9r\") pod \"community-operators-8zg45\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.234422 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.726894 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-9b42b" event={"ID":"761360e0-d8c9-4a45-96a3-c90e19f721fc","Type":"ContainerStarted","Data":"878f4b32fec1ad460081208712cb707fdfe060201df6fa5e3552a9a772840985"} Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.727065 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fcjpj" podUID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" containerName="registry-server" containerID="cri-o://5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb" gracePeriod=2 Nov 24 10:32:44 crc kubenswrapper[4944]: I1124 10:32:44.810614 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8zg45"] Nov 24 10:32:44 crc kubenswrapper[4944]: W1124 10:32:44.830292 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18a76545_dbb6_4e83_ad53_de6924dc985e.slice/crio-76b55cb754480dd487faf5eff9f1bb014390273dabc8ba29b2d7595aa4010c27 WatchSource:0}: Error finding container 76b55cb754480dd487faf5eff9f1bb014390273dabc8ba29b2d7595aa4010c27: Status 404 returned error can't find the container with id 76b55cb754480dd487faf5eff9f1bb014390273dabc8ba29b2d7595aa4010c27 Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.136582 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.302754 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-utilities\") pod \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.303095 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t58f7\" (UniqueName: \"kubernetes.io/projected/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-kube-api-access-t58f7\") pod \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.303313 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-catalog-content\") pod \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\" (UID: \"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02\") " Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.304651 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-utilities" (OuterVolumeSpecName: "utilities") pod "3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" (UID: "3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.315531 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-kube-api-access-t58f7" (OuterVolumeSpecName: "kube-api-access-t58f7") pod "3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" (UID: "3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02"). InnerVolumeSpecName "kube-api-access-t58f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.338008 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" (UID: "3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.405423 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t58f7\" (UniqueName: \"kubernetes.io/projected/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-kube-api-access-t58f7\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.405456 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.405465 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.744199 4944 generic.go:334] "Generic (PLEG): container finished" podID="18a76545-dbb6-4e83-ad53-de6924dc985e" containerID="d991c89f7200bb8a0350986c365dcd50364ebaded59465c2ebb622e3d4aa82a1" exitCode=0 Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.744328 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zg45" event={"ID":"18a76545-dbb6-4e83-ad53-de6924dc985e","Type":"ContainerDied","Data":"d991c89f7200bb8a0350986c365dcd50364ebaded59465c2ebb622e3d4aa82a1"} Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.744373 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zg45" event={"ID":"18a76545-dbb6-4e83-ad53-de6924dc985e","Type":"ContainerStarted","Data":"76b55cb754480dd487faf5eff9f1bb014390273dabc8ba29b2d7595aa4010c27"} Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.749343 4944 generic.go:334] "Generic (PLEG): container finished" podID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" containerID="5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb" exitCode=0 Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.749407 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcjpj" event={"ID":"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02","Type":"ContainerDied","Data":"5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb"} Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.749455 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcjpj" event={"ID":"3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02","Type":"ContainerDied","Data":"25ee29cc8063d07de7d5e22e6b9a83a26f4236b6c2fdd436cc0328f13b1e244e"} Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.749480 4944 scope.go:117] "RemoveContainer" containerID="5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb" Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.749484 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fcjpj" Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.810544 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fcjpj"] Nov 24 10:32:45 crc kubenswrapper[4944]: I1124 10:32:45.827069 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fcjpj"] Nov 24 10:32:46 crc kubenswrapper[4944]: I1124 10:32:46.290152 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" path="/var/lib/kubelet/pods/3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02/volumes" Nov 24 10:32:47 crc kubenswrapper[4944]: I1124 10:32:47.686943 4944 scope.go:117] "RemoveContainer" containerID="0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a" Nov 24 10:32:47 crc kubenswrapper[4944]: I1124 10:32:47.735674 4944 scope.go:117] "RemoveContainer" containerID="d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558" Nov 24 10:32:47 crc kubenswrapper[4944]: I1124 10:32:47.967065 4944 scope.go:117] "RemoveContainer" containerID="5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb" Nov 24 10:32:47 crc kubenswrapper[4944]: E1124 10:32:47.967500 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb\": container with ID starting with 5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb not found: ID does not exist" containerID="5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb" Nov 24 10:32:47 crc kubenswrapper[4944]: I1124 10:32:47.967546 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb"} err="failed to get container status \"5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb\": rpc error: code = NotFound desc = could not find container \"5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb\": container with ID starting with 5d1535e9e48435096926bf326ba8b35a3554c4e220507b38923c12ef084523fb not found: ID does not exist" Nov 24 10:32:47 crc kubenswrapper[4944]: I1124 10:32:47.967572 4944 scope.go:117] "RemoveContainer" containerID="0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a" Nov 24 10:32:47 crc kubenswrapper[4944]: E1124 10:32:47.968001 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a\": container with ID starting with 0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a not found: ID does not exist" containerID="0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a" Nov 24 10:32:47 crc kubenswrapper[4944]: I1124 10:32:47.968077 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a"} err="failed to get container status \"0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a\": rpc error: code = NotFound desc = could not find container \"0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a\": container with ID starting with 0a69f7e0c9f284cd97e461294520d17e88add5711af914f824bad13e91e5917a not found: ID does not exist" Nov 24 10:32:47 crc kubenswrapper[4944]: I1124 10:32:47.968108 4944 scope.go:117] "RemoveContainer" containerID="d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558" Nov 24 10:32:47 crc kubenswrapper[4944]: E1124 10:32:47.968609 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558\": container with ID starting with d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558 not found: ID does not exist" containerID="d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558" Nov 24 10:32:47 crc kubenswrapper[4944]: I1124 10:32:47.968639 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558"} err="failed to get container status \"d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558\": rpc error: code = NotFound desc = could not find container \"d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558\": container with ID starting with d5e99e7843863475a46b6f78f11bd47b8d149ee2a60fe0c2849d897ad8ea1558 not found: ID does not exist" Nov 24 10:32:48 crc kubenswrapper[4944]: I1124 10:32:48.790001 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zg45" event={"ID":"18a76545-dbb6-4e83-ad53-de6924dc985e","Type":"ContainerStarted","Data":"680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1"} Nov 24 10:32:48 crc kubenswrapper[4944]: I1124 10:32:48.793074 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-9b42b" event={"ID":"761360e0-d8c9-4a45-96a3-c90e19f721fc","Type":"ContainerStarted","Data":"b819bb57042c4fa74f8cc6d42e58401e0eab3372a35f0a778fc7f3ea44dfd42e"} Nov 24 10:32:48 crc kubenswrapper[4944]: I1124 10:32:48.856773 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-9b42b" podStartSLOduration=2.869161631 podStartE2EDuration="6.856743764s" podCreationTimestamp="2025-11-24 10:32:42 +0000 UTC" firstStartedPulling="2025-11-24 10:32:43.767808653 +0000 UTC m=+6024.302249115" lastFinishedPulling="2025-11-24 10:32:47.755390786 +0000 UTC m=+6028.289831248" observedRunningTime="2025-11-24 10:32:48.851240309 +0000 UTC m=+6029.385680781" watchObservedRunningTime="2025-11-24 10:32:48.856743764 +0000 UTC m=+6029.391184236" Nov 24 10:32:50 crc kubenswrapper[4944]: I1124 10:32:50.811065 4944 generic.go:334] "Generic (PLEG): container finished" podID="761360e0-d8c9-4a45-96a3-c90e19f721fc" containerID="b819bb57042c4fa74f8cc6d42e58401e0eab3372a35f0a778fc7f3ea44dfd42e" exitCode=0 Nov 24 10:32:50 crc kubenswrapper[4944]: I1124 10:32:50.811154 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-9b42b" event={"ID":"761360e0-d8c9-4a45-96a3-c90e19f721fc","Type":"ContainerDied","Data":"b819bb57042c4fa74f8cc6d42e58401e0eab3372a35f0a778fc7f3ea44dfd42e"} Nov 24 10:32:50 crc kubenswrapper[4944]: I1124 10:32:50.813377 4944 generic.go:334] "Generic (PLEG): container finished" podID="18a76545-dbb6-4e83-ad53-de6924dc985e" containerID="680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1" exitCode=0 Nov 24 10:32:50 crc kubenswrapper[4944]: I1124 10:32:50.813440 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zg45" event={"ID":"18a76545-dbb6-4e83-ad53-de6924dc985e","Type":"ContainerDied","Data":"680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1"} Nov 24 10:32:51 crc kubenswrapper[4944]: I1124 10:32:51.826146 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zg45" event={"ID":"18a76545-dbb6-4e83-ad53-de6924dc985e","Type":"ContainerStarted","Data":"e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189"} Nov 24 10:32:51 crc kubenswrapper[4944]: I1124 10:32:51.850714 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8zg45" podStartSLOduration=4.942945519 podStartE2EDuration="8.850693999s" podCreationTimestamp="2025-11-24 10:32:43 +0000 UTC" firstStartedPulling="2025-11-24 10:32:47.296925764 +0000 UTC m=+6027.831366226" lastFinishedPulling="2025-11-24 10:32:51.204674234 +0000 UTC m=+6031.739114706" observedRunningTime="2025-11-24 10:32:51.845168343 +0000 UTC m=+6032.379608805" watchObservedRunningTime="2025-11-24 10:32:51.850693999 +0000 UTC m=+6032.385134461" Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.303888 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.455822 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hl66\" (UniqueName: \"kubernetes.io/projected/761360e0-d8c9-4a45-96a3-c90e19f721fc-kube-api-access-7hl66\") pod \"761360e0-d8c9-4a45-96a3-c90e19f721fc\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.455867 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-config-data\") pod \"761360e0-d8c9-4a45-96a3-c90e19f721fc\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.455894 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-job-config-data\") pod \"761360e0-d8c9-4a45-96a3-c90e19f721fc\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.455992 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-combined-ca-bundle\") pod \"761360e0-d8c9-4a45-96a3-c90e19f721fc\" (UID: \"761360e0-d8c9-4a45-96a3-c90e19f721fc\") " Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.462542 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/761360e0-d8c9-4a45-96a3-c90e19f721fc-kube-api-access-7hl66" (OuterVolumeSpecName: "kube-api-access-7hl66") pod "761360e0-d8c9-4a45-96a3-c90e19f721fc" (UID: "761360e0-d8c9-4a45-96a3-c90e19f721fc"). InnerVolumeSpecName "kube-api-access-7hl66". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.462999 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "761360e0-d8c9-4a45-96a3-c90e19f721fc" (UID: "761360e0-d8c9-4a45-96a3-c90e19f721fc"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.465386 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-config-data" (OuterVolumeSpecName: "config-data") pod "761360e0-d8c9-4a45-96a3-c90e19f721fc" (UID: "761360e0-d8c9-4a45-96a3-c90e19f721fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.507788 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "761360e0-d8c9-4a45-96a3-c90e19f721fc" (UID: "761360e0-d8c9-4a45-96a3-c90e19f721fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.559332 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.559379 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hl66\" (UniqueName: \"kubernetes.io/projected/761360e0-d8c9-4a45-96a3-c90e19f721fc-kube-api-access-7hl66\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.559391 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.559402 4944 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/761360e0-d8c9-4a45-96a3-c90e19f721fc-job-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.836814 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-9b42b" event={"ID":"761360e0-d8c9-4a45-96a3-c90e19f721fc","Type":"ContainerDied","Data":"878f4b32fec1ad460081208712cb707fdfe060201df6fa5e3552a9a772840985"} Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.836859 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="878f4b32fec1ad460081208712cb707fdfe060201df6fa5e3552a9a772840985" Nov 24 10:32:52 crc kubenswrapper[4944]: I1124 10:32:52.836895 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-9b42b" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.708894 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 10:32:53 crc kubenswrapper[4944]: E1124 10:32:53.710840 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" containerName="extract-utilities" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.710863 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" containerName="extract-utilities" Nov 24 10:32:53 crc kubenswrapper[4944]: E1124 10:32:53.710879 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" containerName="registry-server" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.710886 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" containerName="registry-server" Nov 24 10:32:53 crc kubenswrapper[4944]: E1124 10:32:53.710908 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="761360e0-d8c9-4a45-96a3-c90e19f721fc" containerName="manila-db-sync" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.710914 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="761360e0-d8c9-4a45-96a3-c90e19f721fc" containerName="manila-db-sync" Nov 24 10:32:53 crc kubenswrapper[4944]: E1124 10:32:53.710928 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" containerName="extract-content" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.710934 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" containerName="extract-content" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.712652 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="761360e0-d8c9-4a45-96a3-c90e19f721fc" containerName="manila-db-sync" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.712682 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3930c1d8-0a2e-4035-8ded-b6dd2d2a0b02" containerName="registry-server" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.730191 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.757947 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-94x8g" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.758653 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.759504 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.764844 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.769242 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.771229 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.782556 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.788618 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.804807 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb4408f3-0300-4601-b866-94c5ddb146eb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.804843 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vwg8\" (UniqueName: \"kubernetes.io/projected/eb4408f3-0300-4601-b866-94c5ddb146eb-kube-api-access-5vwg8\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.804868 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.804926 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-config-data\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.804946 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-scripts\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.804972 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d23e50d9-f904-4ee1-8ab4-2731ee07e643-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.804992 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-scripts\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.805017 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.805069 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.805092 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-config-data\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.805121 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d23e50d9-f904-4ee1-8ab4-2731ee07e643-ceph\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.805138 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d23e50d9-f904-4ee1-8ab4-2731ee07e643-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.805161 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.805216 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl7jw\" (UniqueName: \"kubernetes.io/projected/d23e50d9-f904-4ee1-8ab4-2731ee07e643-kube-api-access-gl7jw\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.816547 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.827152 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d99d5fb77-wz575"] Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.829673 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.852869 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d99d5fb77-wz575"] Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.904248 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.905938 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.907784 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb4408f3-0300-4601-b866-94c5ddb146eb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.907815 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vwg8\" (UniqueName: \"kubernetes.io/projected/eb4408f3-0300-4601-b866-94c5ddb146eb-kube-api-access-5vwg8\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.907837 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.907871 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-nb\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.907897 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-sb\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.907934 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-config-data\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.907953 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-scripts\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.907976 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d23e50d9-f904-4ee1-8ab4-2731ee07e643-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.907995 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-scripts\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.908021 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.908067 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.908092 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-config-data\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.908113 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-config\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.908131 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-dns-svc\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.908156 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d23e50d9-f904-4ee1-8ab4-2731ee07e643-ceph\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.908172 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d23e50d9-f904-4ee1-8ab4-2731ee07e643-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.908189 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.908223 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt28g\" (UniqueName: \"kubernetes.io/projected/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-kube-api-access-pt28g\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.908248 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl7jw\" (UniqueName: \"kubernetes.io/projected/d23e50d9-f904-4ee1-8ab4-2731ee07e643-kube-api-access-gl7jw\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.908493 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb4408f3-0300-4601-b866-94c5ddb146eb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.907792 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.912670 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.914913 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.915118 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d23e50d9-f904-4ee1-8ab4-2731ee07e643-ceph\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.915261 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d23e50d9-f904-4ee1-8ab4-2731ee07e643-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.917495 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d23e50d9-f904-4ee1-8ab4-2731ee07e643-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.919414 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-config-data\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.921180 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-scripts\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.921782 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.921863 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-config-data\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.922260 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.922765 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d23e50d9-f904-4ee1-8ab4-2731ee07e643-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.923665 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb4408f3-0300-4601-b866-94c5ddb146eb-scripts\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.932009 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vwg8\" (UniqueName: \"kubernetes.io/projected/eb4408f3-0300-4601-b866-94c5ddb146eb-kube-api-access-5vwg8\") pod \"manila-scheduler-0\" (UID: \"eb4408f3-0300-4601-b866-94c5ddb146eb\") " pod="openstack/manila-scheduler-0" Nov 24 10:32:53 crc kubenswrapper[4944]: I1124 10:32:53.935913 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl7jw\" (UniqueName: \"kubernetes.io/projected/d23e50d9-f904-4ee1-8ab4-2731ee07e643-kube-api-access-gl7jw\") pod \"manila-share-share1-0\" (UID: \"d23e50d9-f904-4ee1-8ab4-2731ee07e643\") " pod="openstack/manila-share-share1-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010230 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-etc-machine-id\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010594 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-config-data-custom\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010630 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-config\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010650 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-dns-svc\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010666 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-scripts\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010709 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010734 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt28g\" (UniqueName: \"kubernetes.io/projected/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-kube-api-access-pt28g\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010767 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94zk9\" (UniqueName: \"kubernetes.io/projected/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-kube-api-access-94zk9\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010809 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-nb\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010829 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-config-data\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010851 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-sb\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.010867 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-logs\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.012184 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-dns-svc\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.012192 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-config\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.012239 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-nb\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.012273 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-sb\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.028012 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt28g\" (UniqueName: \"kubernetes.io/projected/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-kube-api-access-pt28g\") pod \"dnsmasq-dns-7d99d5fb77-wz575\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.089819 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.104716 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.112761 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-etc-machine-id\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.112882 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-config-data-custom\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.112946 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-scripts\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.112959 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-etc-machine-id\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.113020 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.113113 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94zk9\" (UniqueName: \"kubernetes.io/projected/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-kube-api-access-94zk9\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.113166 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-config-data\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.113196 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-logs\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.118249 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-logs\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.119088 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-config-data-custom\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.121098 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-config-data\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.128651 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.130681 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-scripts\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.138270 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94zk9\" (UniqueName: \"kubernetes.io/projected/8c8c89d3-ef27-4d5c-b083-20d868aa86a9-kube-api-access-94zk9\") pod \"manila-api-0\" (UID: \"8c8c89d3-ef27-4d5c-b083-20d868aa86a9\") " pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.174014 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.235962 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.236270 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.347082 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.376572 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.722496 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Nov 24 10:32:54 crc kubenswrapper[4944]: W1124 10:32:54.724187 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb4408f3_0300_4601_b866_94c5ddb146eb.slice/crio-3e31e2aaeadaf750ee1d7bf411fa336838e11eac430f2115ed033fe1b7b15558 WatchSource:0}: Error finding container 3e31e2aaeadaf750ee1d7bf411fa336838e11eac430f2115ed033fe1b7b15558: Status 404 returned error can't find the container with id 3e31e2aaeadaf750ee1d7bf411fa336838e11eac430f2115ed033fe1b7b15558 Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.806462 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.862814 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"eb4408f3-0300-4601-b866-94c5ddb146eb","Type":"ContainerStarted","Data":"3e31e2aaeadaf750ee1d7bf411fa336838e11eac430f2115ed033fe1b7b15558"} Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.864259 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d23e50d9-f904-4ee1-8ab4-2731ee07e643","Type":"ContainerStarted","Data":"20b757c7c335b2540b7e270a19390facba2a7ace3f4ca357c591e2f8b55d2565"} Nov 24 10:32:54 crc kubenswrapper[4944]: I1124 10:32:54.880386 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d99d5fb77-wz575"] Nov 24 10:32:55 crc kubenswrapper[4944]: I1124 10:32:55.071291 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Nov 24 10:32:55 crc kubenswrapper[4944]: W1124 10:32:55.079142 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c8c89d3_ef27_4d5c_b083_20d868aa86a9.slice/crio-d8a51fd411d08c3ea16e5fbf0079198277b63b176ce82f460f4c24f8546cc77b WatchSource:0}: Error finding container d8a51fd411d08c3ea16e5fbf0079198277b63b176ce82f460f4c24f8546cc77b: Status 404 returned error can't find the container with id d8a51fd411d08c3ea16e5fbf0079198277b63b176ce82f460f4c24f8546cc77b Nov 24 10:32:55 crc kubenswrapper[4944]: I1124 10:32:55.886721 4944 generic.go:334] "Generic (PLEG): container finished" podID="4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" containerID="daf896dace00283836bfcd6ed0fa4131ebf3fc3cb1686c2de33706e0f59e50e4" exitCode=0 Nov 24 10:32:55 crc kubenswrapper[4944]: I1124 10:32:55.887145 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" event={"ID":"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6","Type":"ContainerDied","Data":"daf896dace00283836bfcd6ed0fa4131ebf3fc3cb1686c2de33706e0f59e50e4"} Nov 24 10:32:55 crc kubenswrapper[4944]: I1124 10:32:55.887181 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" event={"ID":"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6","Type":"ContainerStarted","Data":"b379a05d76ebe241671bc9bfb2d656abd5c8a9008d7855b5df085ad6b9d5eba1"} Nov 24 10:32:55 crc kubenswrapper[4944]: I1124 10:32:55.896520 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8c8c89d3-ef27-4d5c-b083-20d868aa86a9","Type":"ContainerStarted","Data":"d57422eebfee44c97709272d090bcb00ddb6c490820149f74d9e7472c886f028"} Nov 24 10:32:55 crc kubenswrapper[4944]: I1124 10:32:55.896562 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8c8c89d3-ef27-4d5c-b083-20d868aa86a9","Type":"ContainerStarted","Data":"d8a51fd411d08c3ea16e5fbf0079198277b63b176ce82f460f4c24f8546cc77b"} Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.039763 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-g6srq"] Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.047458 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-g6srq"] Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.315215 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fa3778d-7506-407d-b5d7-b32b310c36d9" path="/var/lib/kubelet/pods/8fa3778d-7506-407d-b5d7-b32b310c36d9/volumes" Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.915150 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"8c8c89d3-ef27-4d5c-b083-20d868aa86a9","Type":"ContainerStarted","Data":"abf0422be3a3955598574993c81fa218bd3ff3dc22611bdceac936a64c1fe532"} Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.915793 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.917954 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"eb4408f3-0300-4601-b866-94c5ddb146eb","Type":"ContainerStarted","Data":"9341802aa2011d6d064c17a97b85abd9dcfb492b2df376ff7649340ba8aaa43b"} Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.918028 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"eb4408f3-0300-4601-b866-94c5ddb146eb","Type":"ContainerStarted","Data":"e2054538baa9c646f73f05355cc044df8ae8dddae0d9296965502626fce67c01"} Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.928002 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" event={"ID":"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6","Type":"ContainerStarted","Data":"0d3f325481b0db6a0fb974381abfa1af0d973fc3fb9f9963c3327bcbe395aa7a"} Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.928186 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.953903 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.953878283 podStartE2EDuration="3.953878283s" podCreationTimestamp="2025-11-24 10:32:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:32:56.943472152 +0000 UTC m=+6037.477912614" watchObservedRunningTime="2025-11-24 10:32:56.953878283 +0000 UTC m=+6037.488318745" Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.971760 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.364116069 podStartE2EDuration="3.971740542s" podCreationTimestamp="2025-11-24 10:32:53 +0000 UTC" firstStartedPulling="2025-11-24 10:32:54.727648029 +0000 UTC m=+6035.262088491" lastFinishedPulling="2025-11-24 10:32:55.335272502 +0000 UTC m=+6035.869712964" observedRunningTime="2025-11-24 10:32:56.971175774 +0000 UTC m=+6037.505616236" watchObservedRunningTime="2025-11-24 10:32:56.971740542 +0000 UTC m=+6037.506181004" Nov 24 10:32:56 crc kubenswrapper[4944]: I1124 10:32:56.997593 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" podStartSLOduration=3.997560075 podStartE2EDuration="3.997560075s" podCreationTimestamp="2025-11-24 10:32:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:32:56.987607107 +0000 UTC m=+6037.522047569" watchObservedRunningTime="2025-11-24 10:32:56.997560075 +0000 UTC m=+6037.532000537" Nov 24 10:32:57 crc kubenswrapper[4944]: I1124 10:32:57.702997 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:32:57 crc kubenswrapper[4944]: I1124 10:32:57.703640 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="ceilometer-central-agent" containerID="cri-o://51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2" gracePeriod=30 Nov 24 10:32:57 crc kubenswrapper[4944]: I1124 10:32:57.703819 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="sg-core" containerID="cri-o://0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0" gracePeriod=30 Nov 24 10:32:57 crc kubenswrapper[4944]: I1124 10:32:57.703870 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="ceilometer-notification-agent" containerID="cri-o://4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993" gracePeriod=30 Nov 24 10:32:57 crc kubenswrapper[4944]: I1124 10:32:57.703819 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="proxy-httpd" containerID="cri-o://849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d" gracePeriod=30 Nov 24 10:32:57 crc kubenswrapper[4944]: I1124 10:32:57.717378 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.1.138:3000/\": EOF" Nov 24 10:32:57 crc kubenswrapper[4944]: I1124 10:32:57.941322 4944 generic.go:334] "Generic (PLEG): container finished" podID="41b72775-aac7-4695-bd24-de5904aeebdb" containerID="849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d" exitCode=0 Nov 24 10:32:57 crc kubenswrapper[4944]: I1124 10:32:57.942311 4944 generic.go:334] "Generic (PLEG): container finished" podID="41b72775-aac7-4695-bd24-de5904aeebdb" containerID="0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0" exitCode=2 Nov 24 10:32:57 crc kubenswrapper[4944]: I1124 10:32:57.941541 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41b72775-aac7-4695-bd24-de5904aeebdb","Type":"ContainerDied","Data":"849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d"} Nov 24 10:32:57 crc kubenswrapper[4944]: I1124 10:32:57.943688 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41b72775-aac7-4695-bd24-de5904aeebdb","Type":"ContainerDied","Data":"0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0"} Nov 24 10:32:58 crc kubenswrapper[4944]: I1124 10:32:58.957768 4944 generic.go:334] "Generic (PLEG): container finished" podID="41b72775-aac7-4695-bd24-de5904aeebdb" containerID="51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2" exitCode=0 Nov 24 10:32:58 crc kubenswrapper[4944]: I1124 10:32:58.957833 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41b72775-aac7-4695-bd24-de5904aeebdb","Type":"ContainerDied","Data":"51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2"} Nov 24 10:33:01 crc kubenswrapper[4944]: I1124 10:33:01.995155 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d23e50d9-f904-4ee1-8ab4-2731ee07e643","Type":"ContainerStarted","Data":"d37bfe50c8aec040fe3fc75db52ea3a9b102241fbe1bbe276d2f75226acc3ff5"} Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.622485 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.715688 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-config-data\") pod \"41b72775-aac7-4695-bd24-de5904aeebdb\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.715781 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-run-httpd\") pod \"41b72775-aac7-4695-bd24-de5904aeebdb\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.715891 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-log-httpd\") pod \"41b72775-aac7-4695-bd24-de5904aeebdb\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.715936 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-sg-core-conf-yaml\") pod \"41b72775-aac7-4695-bd24-de5904aeebdb\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.715982 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-combined-ca-bundle\") pod \"41b72775-aac7-4695-bd24-de5904aeebdb\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.716068 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bk526\" (UniqueName: \"kubernetes.io/projected/41b72775-aac7-4695-bd24-de5904aeebdb-kube-api-access-bk526\") pod \"41b72775-aac7-4695-bd24-de5904aeebdb\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.716099 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-scripts\") pod \"41b72775-aac7-4695-bd24-de5904aeebdb\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.716371 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "41b72775-aac7-4695-bd24-de5904aeebdb" (UID: "41b72775-aac7-4695-bd24-de5904aeebdb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.716796 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "41b72775-aac7-4695-bd24-de5904aeebdb" (UID: "41b72775-aac7-4695-bd24-de5904aeebdb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.717322 4944 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.717341 4944 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41b72775-aac7-4695-bd24-de5904aeebdb-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.722831 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41b72775-aac7-4695-bd24-de5904aeebdb-kube-api-access-bk526" (OuterVolumeSpecName: "kube-api-access-bk526") pod "41b72775-aac7-4695-bd24-de5904aeebdb" (UID: "41b72775-aac7-4695-bd24-de5904aeebdb"). InnerVolumeSpecName "kube-api-access-bk526". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.722956 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-scripts" (OuterVolumeSpecName: "scripts") pod "41b72775-aac7-4695-bd24-de5904aeebdb" (UID: "41b72775-aac7-4695-bd24-de5904aeebdb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.761264 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "41b72775-aac7-4695-bd24-de5904aeebdb" (UID: "41b72775-aac7-4695-bd24-de5904aeebdb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.818414 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41b72775-aac7-4695-bd24-de5904aeebdb" (UID: "41b72775-aac7-4695-bd24-de5904aeebdb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.819137 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-combined-ca-bundle\") pod \"41b72775-aac7-4695-bd24-de5904aeebdb\" (UID: \"41b72775-aac7-4695-bd24-de5904aeebdb\") " Nov 24 10:33:02 crc kubenswrapper[4944]: W1124 10:33:02.819687 4944 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/41b72775-aac7-4695-bd24-de5904aeebdb/volumes/kubernetes.io~secret/combined-ca-bundle Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.819728 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41b72775-aac7-4695-bd24-de5904aeebdb" (UID: "41b72775-aac7-4695-bd24-de5904aeebdb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.820425 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.820455 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bk526\" (UniqueName: \"kubernetes.io/projected/41b72775-aac7-4695-bd24-de5904aeebdb-kube-api-access-bk526\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.820470 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.820482 4944 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.839592 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-config-data" (OuterVolumeSpecName: "config-data") pod "41b72775-aac7-4695-bd24-de5904aeebdb" (UID: "41b72775-aac7-4695-bd24-de5904aeebdb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:33:02 crc kubenswrapper[4944]: I1124 10:33:02.922589 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41b72775-aac7-4695-bd24-de5904aeebdb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.007170 4944 generic.go:334] "Generic (PLEG): container finished" podID="41b72775-aac7-4695-bd24-de5904aeebdb" containerID="4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993" exitCode=0 Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.007198 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.007247 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41b72775-aac7-4695-bd24-de5904aeebdb","Type":"ContainerDied","Data":"4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993"} Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.007287 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41b72775-aac7-4695-bd24-de5904aeebdb","Type":"ContainerDied","Data":"783fe64ac1dea47436a8a689dd5fdaab4270374369a6a6276f6ff318458188d3"} Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.007358 4944 scope.go:117] "RemoveContainer" containerID="849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.011955 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d23e50d9-f904-4ee1-8ab4-2731ee07e643","Type":"ContainerStarted","Data":"57f872f30e9bd8127ead8a47724555fc43c6bc06bf9279862be940891e5dda40"} Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.035610 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.5499465580000003 podStartE2EDuration="10.035592263s" podCreationTimestamp="2025-11-24 10:32:53 +0000 UTC" firstStartedPulling="2025-11-24 10:32:54.807720299 +0000 UTC m=+6035.342160761" lastFinishedPulling="2025-11-24 10:33:01.293366004 +0000 UTC m=+6041.827806466" observedRunningTime="2025-11-24 10:33:03.031036028 +0000 UTC m=+6043.565476500" watchObservedRunningTime="2025-11-24 10:33:03.035592263 +0000 UTC m=+6043.570032725" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.056475 4944 scope.go:117] "RemoveContainer" containerID="0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.082783 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.095321 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.108157 4944 scope.go:117] "RemoveContainer" containerID="4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.109182 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:33:03 crc kubenswrapper[4944]: E1124 10:33:03.109856 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="sg-core" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.109954 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="sg-core" Nov 24 10:33:03 crc kubenswrapper[4944]: E1124 10:33:03.110038 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="ceilometer-central-agent" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.110137 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="ceilometer-central-agent" Nov 24 10:33:03 crc kubenswrapper[4944]: E1124 10:33:03.110237 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="proxy-httpd" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.110311 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="proxy-httpd" Nov 24 10:33:03 crc kubenswrapper[4944]: E1124 10:33:03.110417 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="ceilometer-notification-agent" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.110554 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="ceilometer-notification-agent" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.110980 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="proxy-httpd" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.111110 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="sg-core" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.111205 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="ceilometer-central-agent" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.111313 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" containerName="ceilometer-notification-agent" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.114006 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.116751 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.116936 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.118715 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.137327 4944 scope.go:117] "RemoveContainer" containerID="51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.156636 4944 scope.go:117] "RemoveContainer" containerID="849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d" Nov 24 10:33:03 crc kubenswrapper[4944]: E1124 10:33:03.157038 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d\": container with ID starting with 849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d not found: ID does not exist" containerID="849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.157086 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d"} err="failed to get container status \"849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d\": rpc error: code = NotFound desc = could not find container \"849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d\": container with ID starting with 849459a7d5b27ecc86065897c52a0a73cbba17030530588361b40e091208fb7d not found: ID does not exist" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.157106 4944 scope.go:117] "RemoveContainer" containerID="0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0" Nov 24 10:33:03 crc kubenswrapper[4944]: E1124 10:33:03.158118 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0\": container with ID starting with 0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0 not found: ID does not exist" containerID="0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.158265 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0"} err="failed to get container status \"0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0\": rpc error: code = NotFound desc = could not find container \"0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0\": container with ID starting with 0755c935667fc380107bff3b6d29a252ba88b6b8a52d604232bd25a5eb3c20d0 not found: ID does not exist" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.158371 4944 scope.go:117] "RemoveContainer" containerID="4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993" Nov 24 10:33:03 crc kubenswrapper[4944]: E1124 10:33:03.158827 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993\": container with ID starting with 4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993 not found: ID does not exist" containerID="4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.158910 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993"} err="failed to get container status \"4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993\": rpc error: code = NotFound desc = could not find container \"4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993\": container with ID starting with 4b2c36a1ad30abf4466c4daf8ccc8b92b6ef79c9c70d5575d9c4363dca0aa993 not found: ID does not exist" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.158975 4944 scope.go:117] "RemoveContainer" containerID="51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2" Nov 24 10:33:03 crc kubenswrapper[4944]: E1124 10:33:03.159330 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2\": container with ID starting with 51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2 not found: ID does not exist" containerID="51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.159359 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2"} err="failed to get container status \"51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2\": rpc error: code = NotFound desc = could not find container \"51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2\": container with ID starting with 51923fcdae405e5899f2e36ea9fd07155233f0feebf6129afd5de8d7cbd463d2 not found: ID does not exist" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.231546 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d48n\" (UniqueName: \"kubernetes.io/projected/b4be8251-9473-4b22-9504-cfde2c413f1c-kube-api-access-8d48n\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.231952 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-config-data\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.232205 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-scripts\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.232342 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.232441 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-run-httpd\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.232549 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.232627 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-log-httpd\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.337878 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d48n\" (UniqueName: \"kubernetes.io/projected/b4be8251-9473-4b22-9504-cfde2c413f1c-kube-api-access-8d48n\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.338528 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-config-data\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.339366 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-scripts\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.339975 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.340825 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-run-httpd\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.341333 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-run-httpd\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.340998 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.341435 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-log-httpd\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.341924 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-log-httpd\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.344806 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.345324 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.345561 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-scripts\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.347107 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-config-data\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.362682 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d48n\" (UniqueName: \"kubernetes.io/projected/b4be8251-9473-4b22-9504-cfde2c413f1c-kube-api-access-8d48n\") pod \"ceilometer-0\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.441090 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:33:03 crc kubenswrapper[4944]: I1124 10:33:03.911732 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:33:03 crc kubenswrapper[4944]: W1124 10:33:03.913027 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4be8251_9473_4b22_9504_cfde2c413f1c.slice/crio-e0810b1a32c299f214ce775a2df5755be477adb6f651d7d93cb0b519fa8af3a3 WatchSource:0}: Error finding container e0810b1a32c299f214ce775a2df5755be477adb6f651d7d93cb0b519fa8af3a3: Status 404 returned error can't find the container with id e0810b1a32c299f214ce775a2df5755be477adb6f651d7d93cb0b519fa8af3a3 Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.022585 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4be8251-9473-4b22-9504-cfde2c413f1c","Type":"ContainerStarted","Data":"e0810b1a32c299f214ce775a2df5755be477adb6f651d7d93cb0b519fa8af3a3"} Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.091002 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.105680 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.176227 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.307367 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41b72775-aac7-4695-bd24-de5904aeebdb" path="/var/lib/kubelet/pods/41b72775-aac7-4695-bd24-de5904aeebdb/volumes" Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.308453 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b4c94945c-2ck29"] Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.308707 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" podUID="2d60e5a3-02ae-4c67-8e12-2b0a988cc889" containerName="dnsmasq-dns" containerID="cri-o://cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b" gracePeriod=10 Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.323512 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.397370 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8zg45"] Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.902078 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.982612 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-nb\") pod \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.982804 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-dns-svc\") pod \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.982978 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-sb\") pod \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.983109 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-config\") pod \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.983278 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjs5d\" (UniqueName: \"kubernetes.io/projected/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-kube-api-access-gjs5d\") pod \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\" (UID: \"2d60e5a3-02ae-4c67-8e12-2b0a988cc889\") " Nov 24 10:33:04 crc kubenswrapper[4944]: I1124 10:33:04.992300 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-kube-api-access-gjs5d" (OuterVolumeSpecName: "kube-api-access-gjs5d") pod "2d60e5a3-02ae-4c67-8e12-2b0a988cc889" (UID: "2d60e5a3-02ae-4c67-8e12-2b0a988cc889"). InnerVolumeSpecName "kube-api-access-gjs5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.041757 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2d60e5a3-02ae-4c67-8e12-2b0a988cc889" (UID: "2d60e5a3-02ae-4c67-8e12-2b0a988cc889"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.042033 4944 generic.go:334] "Generic (PLEG): container finished" podID="2d60e5a3-02ae-4c67-8e12-2b0a988cc889" containerID="cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b" exitCode=0 Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.042107 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.042128 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" event={"ID":"2d60e5a3-02ae-4c67-8e12-2b0a988cc889","Type":"ContainerDied","Data":"cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b"} Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.042165 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b4c94945c-2ck29" event={"ID":"2d60e5a3-02ae-4c67-8e12-2b0a988cc889","Type":"ContainerDied","Data":"724132bf0945d8c7310814cfdfb0930aea4967e6b934b822ce5cd27f3db4b63a"} Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.042189 4944 scope.go:117] "RemoveContainer" containerID="cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.044356 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8zg45" podUID="18a76545-dbb6-4e83-ad53-de6924dc985e" containerName="registry-server" containerID="cri-o://e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189" gracePeriod=2 Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.044621 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4be8251-9473-4b22-9504-cfde2c413f1c","Type":"ContainerStarted","Data":"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2"} Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.045218 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2d60e5a3-02ae-4c67-8e12-2b0a988cc889" (UID: "2d60e5a3-02ae-4c67-8e12-2b0a988cc889"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.053725 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-config" (OuterVolumeSpecName: "config") pod "2d60e5a3-02ae-4c67-8e12-2b0a988cc889" (UID: "2d60e5a3-02ae-4c67-8e12-2b0a988cc889"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.062950 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2d60e5a3-02ae-4c67-8e12-2b0a988cc889" (UID: "2d60e5a3-02ae-4c67-8e12-2b0a988cc889"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.074510 4944 scope.go:117] "RemoveContainer" containerID="5a34dfb2847fe4bd8fc30d2886ce66b670b9581eaf95d0ce38a9221962e3f02d" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.086159 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.086203 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.086213 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjs5d\" (UniqueName: \"kubernetes.io/projected/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-kube-api-access-gjs5d\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.086224 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.086232 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d60e5a3-02ae-4c67-8e12-2b0a988cc889-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.101862 4944 scope.go:117] "RemoveContainer" containerID="cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b" Nov 24 10:33:05 crc kubenswrapper[4944]: E1124 10:33:05.102386 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b\": container with ID starting with cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b not found: ID does not exist" containerID="cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.102434 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b"} err="failed to get container status \"cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b\": rpc error: code = NotFound desc = could not find container \"cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b\": container with ID starting with cc8736f0698cc5c098557a27ea39fb2df9e57470e2123d7159dfe5a4154e8f1b not found: ID does not exist" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.102463 4944 scope.go:117] "RemoveContainer" containerID="5a34dfb2847fe4bd8fc30d2886ce66b670b9581eaf95d0ce38a9221962e3f02d" Nov 24 10:33:05 crc kubenswrapper[4944]: E1124 10:33:05.102795 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a34dfb2847fe4bd8fc30d2886ce66b670b9581eaf95d0ce38a9221962e3f02d\": container with ID starting with 5a34dfb2847fe4bd8fc30d2886ce66b670b9581eaf95d0ce38a9221962e3f02d not found: ID does not exist" containerID="5a34dfb2847fe4bd8fc30d2886ce66b670b9581eaf95d0ce38a9221962e3f02d" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.102835 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a34dfb2847fe4bd8fc30d2886ce66b670b9581eaf95d0ce38a9221962e3f02d"} err="failed to get container status \"5a34dfb2847fe4bd8fc30d2886ce66b670b9581eaf95d0ce38a9221962e3f02d\": rpc error: code = NotFound desc = could not find container \"5a34dfb2847fe4bd8fc30d2886ce66b670b9581eaf95d0ce38a9221962e3f02d\": container with ID starting with 5a34dfb2847fe4bd8fc30d2886ce66b670b9581eaf95d0ce38a9221962e3f02d not found: ID does not exist" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.390683 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b4c94945c-2ck29"] Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.414609 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b4c94945c-2ck29"] Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.644271 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.728760 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwq9r\" (UniqueName: \"kubernetes.io/projected/18a76545-dbb6-4e83-ad53-de6924dc985e-kube-api-access-fwq9r\") pod \"18a76545-dbb6-4e83-ad53-de6924dc985e\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.729308 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-utilities\") pod \"18a76545-dbb6-4e83-ad53-de6924dc985e\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.729430 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-catalog-content\") pod \"18a76545-dbb6-4e83-ad53-de6924dc985e\" (UID: \"18a76545-dbb6-4e83-ad53-de6924dc985e\") " Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.729802 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-utilities" (OuterVolumeSpecName: "utilities") pod "18a76545-dbb6-4e83-ad53-de6924dc985e" (UID: "18a76545-dbb6-4e83-ad53-de6924dc985e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.730535 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.748397 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a76545-dbb6-4e83-ad53-de6924dc985e-kube-api-access-fwq9r" (OuterVolumeSpecName: "kube-api-access-fwq9r") pod "18a76545-dbb6-4e83-ad53-de6924dc985e" (UID: "18a76545-dbb6-4e83-ad53-de6924dc985e"). InnerVolumeSpecName "kube-api-access-fwq9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.785024 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18a76545-dbb6-4e83-ad53-de6924dc985e" (UID: "18a76545-dbb6-4e83-ad53-de6924dc985e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.833158 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18a76545-dbb6-4e83-ad53-de6924dc985e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:05 crc kubenswrapper[4944]: I1124 10:33:05.833199 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwq9r\" (UniqueName: \"kubernetes.io/projected/18a76545-dbb6-4e83-ad53-de6924dc985e-kube-api-access-fwq9r\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.077482 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4be8251-9473-4b22-9504-cfde2c413f1c","Type":"ContainerStarted","Data":"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95"} Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.077542 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4be8251-9473-4b22-9504-cfde2c413f1c","Type":"ContainerStarted","Data":"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d"} Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.082978 4944 generic.go:334] "Generic (PLEG): container finished" podID="18a76545-dbb6-4e83-ad53-de6924dc985e" containerID="e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189" exitCode=0 Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.083132 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8zg45" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.083155 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zg45" event={"ID":"18a76545-dbb6-4e83-ad53-de6924dc985e","Type":"ContainerDied","Data":"e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189"} Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.083187 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8zg45" event={"ID":"18a76545-dbb6-4e83-ad53-de6924dc985e","Type":"ContainerDied","Data":"76b55cb754480dd487faf5eff9f1bb014390273dabc8ba29b2d7595aa4010c27"} Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.083205 4944 scope.go:117] "RemoveContainer" containerID="e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.125261 4944 scope.go:117] "RemoveContainer" containerID="680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.128609 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8zg45"] Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.149604 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8zg45"] Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.168846 4944 scope.go:117] "RemoveContainer" containerID="d991c89f7200bb8a0350986c365dcd50364ebaded59465c2ebb622e3d4aa82a1" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.193764 4944 scope.go:117] "RemoveContainer" containerID="e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189" Nov 24 10:33:06 crc kubenswrapper[4944]: E1124 10:33:06.194268 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189\": container with ID starting with e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189 not found: ID does not exist" containerID="e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.194302 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189"} err="failed to get container status \"e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189\": rpc error: code = NotFound desc = could not find container \"e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189\": container with ID starting with e8ebdda99c300387d2ebf233604803a0508516111bb9fb1eed5928db8396f189 not found: ID does not exist" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.194322 4944 scope.go:117] "RemoveContainer" containerID="680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1" Nov 24 10:33:06 crc kubenswrapper[4944]: E1124 10:33:06.194626 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1\": container with ID starting with 680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1 not found: ID does not exist" containerID="680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.194643 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1"} err="failed to get container status \"680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1\": rpc error: code = NotFound desc = could not find container \"680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1\": container with ID starting with 680f3860b38fb4e6a0dae7df1e79b696ce4551663e514d6bf3bd9ff746940ad1 not found: ID does not exist" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.194655 4944 scope.go:117] "RemoveContainer" containerID="d991c89f7200bb8a0350986c365dcd50364ebaded59465c2ebb622e3d4aa82a1" Nov 24 10:33:06 crc kubenswrapper[4944]: E1124 10:33:06.194942 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d991c89f7200bb8a0350986c365dcd50364ebaded59465c2ebb622e3d4aa82a1\": container with ID starting with d991c89f7200bb8a0350986c365dcd50364ebaded59465c2ebb622e3d4aa82a1 not found: ID does not exist" containerID="d991c89f7200bb8a0350986c365dcd50364ebaded59465c2ebb622e3d4aa82a1" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.195003 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d991c89f7200bb8a0350986c365dcd50364ebaded59465c2ebb622e3d4aa82a1"} err="failed to get container status \"d991c89f7200bb8a0350986c365dcd50364ebaded59465c2ebb622e3d4aa82a1\": rpc error: code = NotFound desc = could not find container \"d991c89f7200bb8a0350986c365dcd50364ebaded59465c2ebb622e3d4aa82a1\": container with ID starting with d991c89f7200bb8a0350986c365dcd50364ebaded59465c2ebb622e3d4aa82a1 not found: ID does not exist" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.289830 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18a76545-dbb6-4e83-ad53-de6924dc985e" path="/var/lib/kubelet/pods/18a76545-dbb6-4e83-ad53-de6924dc985e/volumes" Nov 24 10:33:06 crc kubenswrapper[4944]: I1124 10:33:06.290655 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d60e5a3-02ae-4c67-8e12-2b0a988cc889" path="/var/lib/kubelet/pods/2d60e5a3-02ae-4c67-8e12-2b0a988cc889/volumes" Nov 24 10:33:07 crc kubenswrapper[4944]: I1124 10:33:07.910201 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:33:08 crc kubenswrapper[4944]: I1124 10:33:08.112834 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4be8251-9473-4b22-9504-cfde2c413f1c","Type":"ContainerStarted","Data":"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0"} Nov 24 10:33:08 crc kubenswrapper[4944]: I1124 10:33:08.113013 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 10:33:08 crc kubenswrapper[4944]: I1124 10:33:08.149618 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.06820347 podStartE2EDuration="5.149598231s" podCreationTimestamp="2025-11-24 10:33:03 +0000 UTC" firstStartedPulling="2025-11-24 10:33:03.915440756 +0000 UTC m=+6044.449881218" lastFinishedPulling="2025-11-24 10:33:06.996835517 +0000 UTC m=+6047.531275979" observedRunningTime="2025-11-24 10:33:08.132736914 +0000 UTC m=+6048.667177416" watchObservedRunningTime="2025-11-24 10:33:08.149598231 +0000 UTC m=+6048.684038683" Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.125006 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="ceilometer-central-agent" containerID="cri-o://dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2" gracePeriod=30 Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.125109 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="ceilometer-notification-agent" containerID="cri-o://57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d" gracePeriod=30 Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.125183 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="proxy-httpd" containerID="cri-o://f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0" gracePeriod=30 Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.125224 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="sg-core" containerID="cri-o://88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95" gracePeriod=30 Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.790122 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.919464 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-combined-ca-bundle\") pod \"b4be8251-9473-4b22-9504-cfde2c413f1c\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.919551 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-scripts\") pod \"b4be8251-9473-4b22-9504-cfde2c413f1c\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.919618 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-run-httpd\") pod \"b4be8251-9473-4b22-9504-cfde2c413f1c\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.919648 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-sg-core-conf-yaml\") pod \"b4be8251-9473-4b22-9504-cfde2c413f1c\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.919791 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d48n\" (UniqueName: \"kubernetes.io/projected/b4be8251-9473-4b22-9504-cfde2c413f1c-kube-api-access-8d48n\") pod \"b4be8251-9473-4b22-9504-cfde2c413f1c\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.919860 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-config-data\") pod \"b4be8251-9473-4b22-9504-cfde2c413f1c\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.919933 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-log-httpd\") pod \"b4be8251-9473-4b22-9504-cfde2c413f1c\" (UID: \"b4be8251-9473-4b22-9504-cfde2c413f1c\") " Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.920375 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b4be8251-9473-4b22-9504-cfde2c413f1c" (UID: "b4be8251-9473-4b22-9504-cfde2c413f1c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.920535 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b4be8251-9473-4b22-9504-cfde2c413f1c" (UID: "b4be8251-9473-4b22-9504-cfde2c413f1c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.924760 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4be8251-9473-4b22-9504-cfde2c413f1c-kube-api-access-8d48n" (OuterVolumeSpecName: "kube-api-access-8d48n") pod "b4be8251-9473-4b22-9504-cfde2c413f1c" (UID: "b4be8251-9473-4b22-9504-cfde2c413f1c"). InnerVolumeSpecName "kube-api-access-8d48n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.925271 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-scripts" (OuterVolumeSpecName: "scripts") pod "b4be8251-9473-4b22-9504-cfde2c413f1c" (UID: "b4be8251-9473-4b22-9504-cfde2c413f1c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.948459 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b4be8251-9473-4b22-9504-cfde2c413f1c" (UID: "b4be8251-9473-4b22-9504-cfde2c413f1c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:33:09 crc kubenswrapper[4944]: I1124 10:33:09.998180 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4be8251-9473-4b22-9504-cfde2c413f1c" (UID: "b4be8251-9473-4b22-9504-cfde2c413f1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.018142 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-config-data" (OuterVolumeSpecName: "config-data") pod "b4be8251-9473-4b22-9504-cfde2c413f1c" (UID: "b4be8251-9473-4b22-9504-cfde2c413f1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.022645 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d48n\" (UniqueName: \"kubernetes.io/projected/b4be8251-9473-4b22-9504-cfde2c413f1c-kube-api-access-8d48n\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.022673 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.022683 4944 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.022692 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.022701 4944 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.022710 4944 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4be8251-9473-4b22-9504-cfde2c413f1c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.022720 4944 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4be8251-9473-4b22-9504-cfde2c413f1c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.136689 4944 generic.go:334] "Generic (PLEG): container finished" podID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerID="f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0" exitCode=0 Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.136729 4944 generic.go:334] "Generic (PLEG): container finished" podID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerID="88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95" exitCode=2 Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.136740 4944 generic.go:334] "Generic (PLEG): container finished" podID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerID="57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d" exitCode=0 Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.136729 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4be8251-9473-4b22-9504-cfde2c413f1c","Type":"ContainerDied","Data":"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0"} Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.136792 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4be8251-9473-4b22-9504-cfde2c413f1c","Type":"ContainerDied","Data":"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95"} Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.136808 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4be8251-9473-4b22-9504-cfde2c413f1c","Type":"ContainerDied","Data":"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d"} Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.136810 4944 scope.go:117] "RemoveContainer" containerID="f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.136817 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4be8251-9473-4b22-9504-cfde2c413f1c","Type":"ContainerDied","Data":"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2"} Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.136749 4944 generic.go:334] "Generic (PLEG): container finished" podID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerID="dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2" exitCode=0 Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.136921 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4be8251-9473-4b22-9504-cfde2c413f1c","Type":"ContainerDied","Data":"e0810b1a32c299f214ce775a2df5755be477adb6f651d7d93cb0b519fa8af3a3"} Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.138205 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.162217 4944 scope.go:117] "RemoveContainer" containerID="88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.171497 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.179099 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.204289 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.204796 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d60e5a3-02ae-4c67-8e12-2b0a988cc889" containerName="init" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.204816 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d60e5a3-02ae-4c67-8e12-2b0a988cc889" containerName="init" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.204840 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d60e5a3-02ae-4c67-8e12-2b0a988cc889" containerName="dnsmasq-dns" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.204846 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d60e5a3-02ae-4c67-8e12-2b0a988cc889" containerName="dnsmasq-dns" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.204858 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a76545-dbb6-4e83-ad53-de6924dc985e" containerName="registry-server" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.204897 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a76545-dbb6-4e83-ad53-de6924dc985e" containerName="registry-server" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.204913 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="sg-core" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.204919 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="sg-core" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.204928 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a76545-dbb6-4e83-ad53-de6924dc985e" containerName="extract-utilities" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.204936 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a76545-dbb6-4e83-ad53-de6924dc985e" containerName="extract-utilities" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.204957 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="ceilometer-notification-agent" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.204966 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="ceilometer-notification-agent" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.204982 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a76545-dbb6-4e83-ad53-de6924dc985e" containerName="extract-content" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.204988 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a76545-dbb6-4e83-ad53-de6924dc985e" containerName="extract-content" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.205001 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="ceilometer-central-agent" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.205007 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="ceilometer-central-agent" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.205018 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="proxy-httpd" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.205024 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="proxy-httpd" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.205415 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d60e5a3-02ae-4c67-8e12-2b0a988cc889" containerName="dnsmasq-dns" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.205441 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="sg-core" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.205452 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="ceilometer-central-agent" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.205460 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a76545-dbb6-4e83-ad53-de6924dc985e" containerName="registry-server" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.205470 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="ceilometer-notification-agent" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.205483 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" containerName="proxy-httpd" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.207371 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.213873 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.214467 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.226998 4944 scope.go:117] "RemoveContainer" containerID="57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.231192 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.294190 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4be8251-9473-4b22-9504-cfde2c413f1c" path="/var/lib/kubelet/pods/b4be8251-9473-4b22-9504-cfde2c413f1c/volumes" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.312354 4944 scope.go:117] "RemoveContainer" containerID="dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.329836 4944 scope.go:117] "RemoveContainer" containerID="f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.330327 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0\": container with ID starting with f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0 not found: ID does not exist" containerID="f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.330391 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0"} err="failed to get container status \"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0\": rpc error: code = NotFound desc = could not find container \"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0\": container with ID starting with f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.330420 4944 scope.go:117] "RemoveContainer" containerID="88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.330816 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95\": container with ID starting with 88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95 not found: ID does not exist" containerID="88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.330857 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95"} err="failed to get container status \"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95\": rpc error: code = NotFound desc = could not find container \"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95\": container with ID starting with 88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.330881 4944 scope.go:117] "RemoveContainer" containerID="57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.331233 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d\": container with ID starting with 57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d not found: ID does not exist" containerID="57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.331278 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d"} err="failed to get container status \"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d\": rpc error: code = NotFound desc = could not find container \"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d\": container with ID starting with 57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.331296 4944 scope.go:117] "RemoveContainer" containerID="dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.332569 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.332604 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ndtp\" (UniqueName: \"kubernetes.io/projected/ec6f15bc-0236-428d-9609-b1b24f963628-kube-api-access-5ndtp\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.332637 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec6f15bc-0236-428d-9609-b1b24f963628-log-httpd\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.332667 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.332689 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-config-data\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.332734 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec6f15bc-0236-428d-9609-b1b24f963628-run-httpd\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.332754 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-scripts\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: E1124 10:33:10.332941 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2\": container with ID starting with dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2 not found: ID does not exist" containerID="dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.332966 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2"} err="failed to get container status \"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2\": rpc error: code = NotFound desc = could not find container \"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2\": container with ID starting with dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.332982 4944 scope.go:117] "RemoveContainer" containerID="f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.333576 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0"} err="failed to get container status \"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0\": rpc error: code = NotFound desc = could not find container \"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0\": container with ID starting with f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.333597 4944 scope.go:117] "RemoveContainer" containerID="88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.333790 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95"} err="failed to get container status \"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95\": rpc error: code = NotFound desc = could not find container \"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95\": container with ID starting with 88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.333812 4944 scope.go:117] "RemoveContainer" containerID="57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.334152 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d"} err="failed to get container status \"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d\": rpc error: code = NotFound desc = could not find container \"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d\": container with ID starting with 57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.334171 4944 scope.go:117] "RemoveContainer" containerID="dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.334411 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2"} err="failed to get container status \"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2\": rpc error: code = NotFound desc = could not find container \"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2\": container with ID starting with dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.334430 4944 scope.go:117] "RemoveContainer" containerID="f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.334649 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0"} err="failed to get container status \"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0\": rpc error: code = NotFound desc = could not find container \"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0\": container with ID starting with f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.334687 4944 scope.go:117] "RemoveContainer" containerID="88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.334934 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95"} err="failed to get container status \"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95\": rpc error: code = NotFound desc = could not find container \"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95\": container with ID starting with 88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.334960 4944 scope.go:117] "RemoveContainer" containerID="57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.335191 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d"} err="failed to get container status \"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d\": rpc error: code = NotFound desc = could not find container \"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d\": container with ID starting with 57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.335212 4944 scope.go:117] "RemoveContainer" containerID="dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.335411 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2"} err="failed to get container status \"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2\": rpc error: code = NotFound desc = could not find container \"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2\": container with ID starting with dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.335447 4944 scope.go:117] "RemoveContainer" containerID="f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.335676 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0"} err="failed to get container status \"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0\": rpc error: code = NotFound desc = could not find container \"f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0\": container with ID starting with f016cf81e60da3b55cf05cf33bfb9a4fdd733e2982590d5088e8d6de0c8e85c0 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.335696 4944 scope.go:117] "RemoveContainer" containerID="88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.335871 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95"} err="failed to get container status \"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95\": rpc error: code = NotFound desc = could not find container \"88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95\": container with ID starting with 88acd7703b546ed85dc60a5d07a826878ce8255c13ff555f3ca2e5f41d48af95 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.335892 4944 scope.go:117] "RemoveContainer" containerID="57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.336108 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d"} err="failed to get container status \"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d\": rpc error: code = NotFound desc = could not find container \"57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d\": container with ID starting with 57db1a80774f35876bb44fe71900cc8dbc933c5dd9039cb17e5d9c212850102d not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.336125 4944 scope.go:117] "RemoveContainer" containerID="dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.336323 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2"} err="failed to get container status \"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2\": rpc error: code = NotFound desc = could not find container \"dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2\": container with ID starting with dc0695ff5b8e97a3605bca4a94c5c2ea71284a2c1fa3da28e81e66c1baf36ce2 not found: ID does not exist" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.434506 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-config-data\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.434581 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec6f15bc-0236-428d-9609-b1b24f963628-run-httpd\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.434607 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-scripts\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.434741 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.434762 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ndtp\" (UniqueName: \"kubernetes.io/projected/ec6f15bc-0236-428d-9609-b1b24f963628-kube-api-access-5ndtp\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.434791 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec6f15bc-0236-428d-9609-b1b24f963628-log-httpd\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.434820 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.437334 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec6f15bc-0236-428d-9609-b1b24f963628-log-httpd\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.437724 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ec6f15bc-0236-428d-9609-b1b24f963628-run-httpd\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.441000 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.441135 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.441217 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-config-data\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.442274 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec6f15bc-0236-428d-9609-b1b24f963628-scripts\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.453506 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ndtp\" (UniqueName: \"kubernetes.io/projected/ec6f15bc-0236-428d-9609-b1b24f963628-kube-api-access-5ndtp\") pod \"ceilometer-0\" (UID: \"ec6f15bc-0236-428d-9609-b1b24f963628\") " pod="openstack/ceilometer-0" Nov 24 10:33:10 crc kubenswrapper[4944]: I1124 10:33:10.601514 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 10:33:11 crc kubenswrapper[4944]: I1124 10:33:11.061128 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 10:33:11 crc kubenswrapper[4944]: W1124 10:33:11.065351 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec6f15bc_0236_428d_9609_b1b24f963628.slice/crio-5404958f0f535b63ffd108c916c09fe401d5c3dab4685c0f6a9ccb41f718ad1e WatchSource:0}: Error finding container 5404958f0f535b63ffd108c916c09fe401d5c3dab4685c0f6a9ccb41f718ad1e: Status 404 returned error can't find the container with id 5404958f0f535b63ffd108c916c09fe401d5c3dab4685c0f6a9ccb41f718ad1e Nov 24 10:33:11 crc kubenswrapper[4944]: I1124 10:33:11.150267 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec6f15bc-0236-428d-9609-b1b24f963628","Type":"ContainerStarted","Data":"5404958f0f535b63ffd108c916c09fe401d5c3dab4685c0f6a9ccb41f718ad1e"} Nov 24 10:33:12 crc kubenswrapper[4944]: I1124 10:33:12.168349 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec6f15bc-0236-428d-9609-b1b24f963628","Type":"ContainerStarted","Data":"7b24a0b4f2a000c4019b3e2615826894879f016a58ecb92cefba455d25843299"} Nov 24 10:33:13 crc kubenswrapper[4944]: I1124 10:33:13.179089 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec6f15bc-0236-428d-9609-b1b24f963628","Type":"ContainerStarted","Data":"d263c63fce16b6198269ea81bb8d5812b92b6b503fa3ca2cab65c483071c67b7"} Nov 24 10:33:13 crc kubenswrapper[4944]: I1124 10:33:13.179592 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec6f15bc-0236-428d-9609-b1b24f963628","Type":"ContainerStarted","Data":"f4650e6694758d5043eab28b5eb850e36be38555c4c7ac31744069d9eca2882c"} Nov 24 10:33:14 crc kubenswrapper[4944]: I1124 10:33:14.192148 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ec6f15bc-0236-428d-9609-b1b24f963628","Type":"ContainerStarted","Data":"831957093c364eeb89ced0b63c771533aea44039c36df53bfc6e1c3451606432"} Nov 24 10:33:14 crc kubenswrapper[4944]: I1124 10:33:14.192687 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 10:33:14 crc kubenswrapper[4944]: I1124 10:33:14.222609 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.344599681 podStartE2EDuration="4.222580803s" podCreationTimestamp="2025-11-24 10:33:10 +0000 UTC" firstStartedPulling="2025-11-24 10:33:11.068295281 +0000 UTC m=+6051.602735743" lastFinishedPulling="2025-11-24 10:33:13.946276403 +0000 UTC m=+6054.480716865" observedRunningTime="2025-11-24 10:33:14.216998835 +0000 UTC m=+6054.751439317" watchObservedRunningTime="2025-11-24 10:33:14.222580803 +0000 UTC m=+6054.757021265" Nov 24 10:33:15 crc kubenswrapper[4944]: I1124 10:33:15.563246 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Nov 24 10:33:15 crc kubenswrapper[4944]: I1124 10:33:15.711661 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Nov 24 10:33:15 crc kubenswrapper[4944]: I1124 10:33:15.713732 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Nov 24 10:33:33 crc kubenswrapper[4944]: I1124 10:33:33.805232 4944 scope.go:117] "RemoveContainer" containerID="1f07ae27e5791b89cf80fde0600cc7c6e93b782f0e74fb16ca2aeec496f19b2f" Nov 24 10:33:33 crc kubenswrapper[4944]: I1124 10:33:33.868124 4944 scope.go:117] "RemoveContainer" containerID="8c25a2f788ade6a6f856dd1cb31fcc0baa65f54d51b1e64b85890232a5d92ee0" Nov 24 10:33:33 crc kubenswrapper[4944]: I1124 10:33:33.907423 4944 scope.go:117] "RemoveContainer" containerID="f598824cf5db9267ba2299c12065aca660b33723888f50211a17bcf26b4903a6" Nov 24 10:33:33 crc kubenswrapper[4944]: I1124 10:33:33.951625 4944 scope.go:117] "RemoveContainer" containerID="0e722ee180ad0881308285bd522a6caafbf07cdb300f271b5e59c675789ee2e3" Nov 24 10:33:38 crc kubenswrapper[4944]: I1124 10:33:38.049772 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-9dae-account-create-xbfr7"] Nov 24 10:33:38 crc kubenswrapper[4944]: I1124 10:33:38.058721 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-m29ll"] Nov 24 10:33:38 crc kubenswrapper[4944]: I1124 10:33:38.066741 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-9dae-account-create-xbfr7"] Nov 24 10:33:38 crc kubenswrapper[4944]: I1124 10:33:38.076634 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-m29ll"] Nov 24 10:33:38 crc kubenswrapper[4944]: I1124 10:33:38.290474 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e93583a-c60e-4665-8b4a-aad77baf1ac6" path="/var/lib/kubelet/pods/5e93583a-c60e-4665-8b4a-aad77baf1ac6/volumes" Nov 24 10:33:38 crc kubenswrapper[4944]: I1124 10:33:38.291336 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d028bc5d-4e09-41eb-bf7b-78de5e065799" path="/var/lib/kubelet/pods/d028bc5d-4e09-41eb-bf7b-78de5e065799/volumes" Nov 24 10:33:40 crc kubenswrapper[4944]: I1124 10:33:40.613181 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 10:33:45 crc kubenswrapper[4944]: I1124 10:33:45.033086 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-kczbq"] Nov 24 10:33:45 crc kubenswrapper[4944]: I1124 10:33:45.043985 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-kczbq"] Nov 24 10:33:46 crc kubenswrapper[4944]: I1124 10:33:46.289614 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ed874ef-a00f-47b9-a9d8-18e2a8836a8b" path="/var/lib/kubelet/pods/0ed874ef-a00f-47b9-a9d8-18e2a8836a8b/volumes" Nov 24 10:34:01 crc kubenswrapper[4944]: I1124 10:34:01.772864 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dfbdd8fcf-fssr8"] Nov 24 10:34:01 crc kubenswrapper[4944]: I1124 10:34:01.776071 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:01 crc kubenswrapper[4944]: I1124 10:34:01.778115 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Nov 24 10:34:01 crc kubenswrapper[4944]: I1124 10:34:01.783847 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dfbdd8fcf-fssr8"] Nov 24 10:34:01 crc kubenswrapper[4944]: I1124 10:34:01.927028 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-openstack-cell1\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:01 crc kubenswrapper[4944]: I1124 10:34:01.927131 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-nb\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:01 crc kubenswrapper[4944]: I1124 10:34:01.927169 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-dns-svc\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:01 crc kubenswrapper[4944]: I1124 10:34:01.927209 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-sb\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:01 crc kubenswrapper[4944]: I1124 10:34:01.927257 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-config\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:01 crc kubenswrapper[4944]: I1124 10:34:01.927298 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c87xc\" (UniqueName: \"kubernetes.io/projected/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-kube-api-access-c87xc\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.029142 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-sb\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.029264 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-config\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.029354 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c87xc\" (UniqueName: \"kubernetes.io/projected/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-kube-api-access-c87xc\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.029401 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-openstack-cell1\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.029464 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-nb\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.029495 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-dns-svc\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.030419 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-sb\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.030465 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-dns-svc\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.030495 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-openstack-cell1\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.030654 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-config\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.030654 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-nb\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.047619 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c87xc\" (UniqueName: \"kubernetes.io/projected/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-kube-api-access-c87xc\") pod \"dnsmasq-dns-5dfbdd8fcf-fssr8\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.141034 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.592324 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dfbdd8fcf-fssr8"] Nov 24 10:34:02 crc kubenswrapper[4944]: I1124 10:34:02.706998 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" event={"ID":"f6b22277-7b6e-457b-8dd1-4bd9bff48f80","Type":"ContainerStarted","Data":"2c131ac9f4f398bf6ea688ca6784c1db25b929b1b8505a378afe1b5181147053"} Nov 24 10:34:03 crc kubenswrapper[4944]: I1124 10:34:03.718131 4944 generic.go:334] "Generic (PLEG): container finished" podID="f6b22277-7b6e-457b-8dd1-4bd9bff48f80" containerID="20b0bf94f954ad635cbcb67f56a3c379f3b80ce48395a3fc25a491ea781e47e1" exitCode=0 Nov 24 10:34:03 crc kubenswrapper[4944]: I1124 10:34:03.718269 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" event={"ID":"f6b22277-7b6e-457b-8dd1-4bd9bff48f80","Type":"ContainerDied","Data":"20b0bf94f954ad635cbcb67f56a3c379f3b80ce48395a3fc25a491ea781e47e1"} Nov 24 10:34:04 crc kubenswrapper[4944]: I1124 10:34:04.733631 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" event={"ID":"f6b22277-7b6e-457b-8dd1-4bd9bff48f80","Type":"ContainerStarted","Data":"9689017c192dd74e5ffbf3a11cbca2113784a2663d9f70cc917c6207d1d2438c"} Nov 24 10:34:04 crc kubenswrapper[4944]: I1124 10:34:04.734222 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:04 crc kubenswrapper[4944]: I1124 10:34:04.760331 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" podStartSLOduration=3.760300247 podStartE2EDuration="3.760300247s" podCreationTimestamp="2025-11-24 10:34:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:34:04.756290309 +0000 UTC m=+6105.290730801" watchObservedRunningTime="2025-11-24 10:34:04.760300247 +0000 UTC m=+6105.294740709" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.142932 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.214847 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d99d5fb77-wz575"] Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.215113 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" podUID="4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" containerName="dnsmasq-dns" containerID="cri-o://0d3f325481b0db6a0fb974381abfa1af0d973fc3fb9f9963c3327bcbe395aa7a" gracePeriod=10 Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.349128 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b454bb789-4gtgk"] Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.360603 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.372710 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b454bb789-4gtgk"] Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.465506 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-openstack-cell1\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.465643 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-ovsdbserver-sb\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.465722 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-ovsdbserver-nb\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.465749 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-config\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.465801 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bd27\" (UniqueName: \"kubernetes.io/projected/8af65607-21e8-4d80-8d5f-12faa2e45caf-kube-api-access-6bd27\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.465822 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-dns-svc\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.568898 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bd27\" (UniqueName: \"kubernetes.io/projected/8af65607-21e8-4d80-8d5f-12faa2e45caf-kube-api-access-6bd27\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.568935 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-dns-svc\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.569029 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-openstack-cell1\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.569130 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-ovsdbserver-sb\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.569204 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-ovsdbserver-nb\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.569225 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-config\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.570100 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-config\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.571539 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-openstack-cell1\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.571647 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-ovsdbserver-nb\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.571977 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-ovsdbserver-sb\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.573724 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8af65607-21e8-4d80-8d5f-12faa2e45caf-dns-svc\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.612479 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bd27\" (UniqueName: \"kubernetes.io/projected/8af65607-21e8-4d80-8d5f-12faa2e45caf-kube-api-access-6bd27\") pod \"dnsmasq-dns-5b454bb789-4gtgk\" (UID: \"8af65607-21e8-4d80-8d5f-12faa2e45caf\") " pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.708117 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.825324 4944 generic.go:334] "Generic (PLEG): container finished" podID="4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" containerID="0d3f325481b0db6a0fb974381abfa1af0d973fc3fb9f9963c3327bcbe395aa7a" exitCode=0 Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.825369 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" event={"ID":"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6","Type":"ContainerDied","Data":"0d3f325481b0db6a0fb974381abfa1af0d973fc3fb9f9963c3327bcbe395aa7a"} Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.825395 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" event={"ID":"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6","Type":"ContainerDied","Data":"b379a05d76ebe241671bc9bfb2d656abd5c8a9008d7855b5df085ad6b9d5eba1"} Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.825406 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b379a05d76ebe241671bc9bfb2d656abd5c8a9008d7855b5df085ad6b9d5eba1" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.860618 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.983152 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-config\") pod \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.983255 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-dns-svc\") pod \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.983508 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-sb\") pod \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.983558 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-nb\") pod \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.983657 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt28g\" (UniqueName: \"kubernetes.io/projected/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-kube-api-access-pt28g\") pod \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\" (UID: \"4d773f7a-f45b-4bdd-8167-9e7415a7d7b6\") " Nov 24 10:34:12 crc kubenswrapper[4944]: I1124 10:34:12.994301 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-kube-api-access-pt28g" (OuterVolumeSpecName: "kube-api-access-pt28g") pod "4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" (UID: "4d773f7a-f45b-4bdd-8167-9e7415a7d7b6"). InnerVolumeSpecName "kube-api-access-pt28g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.049757 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" (UID: "4d773f7a-f45b-4bdd-8167-9e7415a7d7b6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.070739 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" (UID: "4d773f7a-f45b-4bdd-8167-9e7415a7d7b6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.071115 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-config" (OuterVolumeSpecName: "config") pod "4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" (UID: "4d773f7a-f45b-4bdd-8167-9e7415a7d7b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.086878 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt28g\" (UniqueName: \"kubernetes.io/projected/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-kube-api-access-pt28g\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.086918 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.086927 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.086938 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.101721 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" (UID: "4d773f7a-f45b-4bdd-8167-9e7415a7d7b6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.186586 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b454bb789-4gtgk"] Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.191179 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.834691 4944 generic.go:334] "Generic (PLEG): container finished" podID="8af65607-21e8-4d80-8d5f-12faa2e45caf" containerID="2bf7c61b19de97e7fab7909ceb5ef0490eb87118f7b1df248bfd7b3520d7b6b8" exitCode=0 Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.834739 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" event={"ID":"8af65607-21e8-4d80-8d5f-12faa2e45caf","Type":"ContainerDied","Data":"2bf7c61b19de97e7fab7909ceb5ef0490eb87118f7b1df248bfd7b3520d7b6b8"} Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.834791 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" event={"ID":"8af65607-21e8-4d80-8d5f-12faa2e45caf","Type":"ContainerStarted","Data":"dab115821482ec3f13bb38a4748a8c2377f88db153fce998b311f0888f5932ec"} Nov 24 10:34:13 crc kubenswrapper[4944]: I1124 10:34:13.834800 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d99d5fb77-wz575" Nov 24 10:34:14 crc kubenswrapper[4944]: I1124 10:34:14.034686 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d99d5fb77-wz575"] Nov 24 10:34:14 crc kubenswrapper[4944]: I1124 10:34:14.043630 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d99d5fb77-wz575"] Nov 24 10:34:14 crc kubenswrapper[4944]: I1124 10:34:14.292544 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" path="/var/lib/kubelet/pods/4d773f7a-f45b-4bdd-8167-9e7415a7d7b6/volumes" Nov 24 10:34:14 crc kubenswrapper[4944]: I1124 10:34:14.844820 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" event={"ID":"8af65607-21e8-4d80-8d5f-12faa2e45caf","Type":"ContainerStarted","Data":"e7ae03346e965e1ecc1ad26cde6f9635121d46602af9171bc46697cb6b02ef64"} Nov 24 10:34:14 crc kubenswrapper[4944]: I1124 10:34:14.845324 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:22 crc kubenswrapper[4944]: I1124 10:34:22.709219 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" Nov 24 10:34:22 crc kubenswrapper[4944]: I1124 10:34:22.748302 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b454bb789-4gtgk" podStartSLOduration=10.748271236 podStartE2EDuration="10.748271236s" podCreationTimestamp="2025-11-24 10:34:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 10:34:14.871645088 +0000 UTC m=+6115.406085550" watchObservedRunningTime="2025-11-24 10:34:22.748271236 +0000 UTC m=+6123.282711738" Nov 24 10:34:22 crc kubenswrapper[4944]: I1124 10:34:22.791807 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dfbdd8fcf-fssr8"] Nov 24 10:34:22 crc kubenswrapper[4944]: I1124 10:34:22.792496 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" podUID="f6b22277-7b6e-457b-8dd1-4bd9bff48f80" containerName="dnsmasq-dns" containerID="cri-o://9689017c192dd74e5ffbf3a11cbca2113784a2663d9f70cc917c6207d1d2438c" gracePeriod=10 Nov 24 10:34:22 crc kubenswrapper[4944]: I1124 10:34:22.925374 4944 generic.go:334] "Generic (PLEG): container finished" podID="f6b22277-7b6e-457b-8dd1-4bd9bff48f80" containerID="9689017c192dd74e5ffbf3a11cbca2113784a2663d9f70cc917c6207d1d2438c" exitCode=0 Nov 24 10:34:22 crc kubenswrapper[4944]: I1124 10:34:22.925415 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" event={"ID":"f6b22277-7b6e-457b-8dd1-4bd9bff48f80","Type":"ContainerDied","Data":"9689017c192dd74e5ffbf3a11cbca2113784a2663d9f70cc917c6207d1d2438c"} Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.350440 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.439630 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-nb\") pod \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.439713 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c87xc\" (UniqueName: \"kubernetes.io/projected/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-kube-api-access-c87xc\") pod \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.439754 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-sb\") pod \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.439800 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-config\") pod \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.439850 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-dns-svc\") pod \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.439893 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-openstack-cell1\") pod \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\" (UID: \"f6b22277-7b6e-457b-8dd1-4bd9bff48f80\") " Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.456826 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-kube-api-access-c87xc" (OuterVolumeSpecName: "kube-api-access-c87xc") pod "f6b22277-7b6e-457b-8dd1-4bd9bff48f80" (UID: "f6b22277-7b6e-457b-8dd1-4bd9bff48f80"). InnerVolumeSpecName "kube-api-access-c87xc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.531064 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f6b22277-7b6e-457b-8dd1-4bd9bff48f80" (UID: "f6b22277-7b6e-457b-8dd1-4bd9bff48f80"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.535037 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "f6b22277-7b6e-457b-8dd1-4bd9bff48f80" (UID: "f6b22277-7b6e-457b-8dd1-4bd9bff48f80"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.542800 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c87xc\" (UniqueName: \"kubernetes.io/projected/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-kube-api-access-c87xc\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.542859 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.542870 4944 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-openstack-cell1\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.546271 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f6b22277-7b6e-457b-8dd1-4bd9bff48f80" (UID: "f6b22277-7b6e-457b-8dd1-4bd9bff48f80"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.547532 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f6b22277-7b6e-457b-8dd1-4bd9bff48f80" (UID: "f6b22277-7b6e-457b-8dd1-4bd9bff48f80"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.548712 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.548787 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.550911 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-config" (OuterVolumeSpecName: "config") pod "f6b22277-7b6e-457b-8dd1-4bd9bff48f80" (UID: "f6b22277-7b6e-457b-8dd1-4bd9bff48f80"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.644152 4944 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.644179 4944 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-config\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.644190 4944 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b22277-7b6e-457b-8dd1-4bd9bff48f80-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.934915 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" event={"ID":"f6b22277-7b6e-457b-8dd1-4bd9bff48f80","Type":"ContainerDied","Data":"2c131ac9f4f398bf6ea688ca6784c1db25b929b1b8505a378afe1b5181147053"} Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.934973 4944 scope.go:117] "RemoveContainer" containerID="9689017c192dd74e5ffbf3a11cbca2113784a2663d9f70cc917c6207d1d2438c" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.934974 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfbdd8fcf-fssr8" Nov 24 10:34:23 crc kubenswrapper[4944]: I1124 10:34:23.982250 4944 scope.go:117] "RemoveContainer" containerID="20b0bf94f954ad635cbcb67f56a3c379f3b80ce48395a3fc25a491ea781e47e1" Nov 24 10:34:24 crc kubenswrapper[4944]: I1124 10:34:24.004502 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dfbdd8fcf-fssr8"] Nov 24 10:34:24 crc kubenswrapper[4944]: I1124 10:34:24.019986 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dfbdd8fcf-fssr8"] Nov 24 10:34:24 crc kubenswrapper[4944]: I1124 10:34:24.287534 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6b22277-7b6e-457b-8dd1-4bd9bff48f80" path="/var/lib/kubelet/pods/f6b22277-7b6e-457b-8dd1-4bd9bff48f80/volumes" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.817486 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f"] Nov 24 10:34:33 crc kubenswrapper[4944]: E1124 10:34:33.818770 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" containerName="dnsmasq-dns" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.818793 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" containerName="dnsmasq-dns" Nov 24 10:34:33 crc kubenswrapper[4944]: E1124 10:34:33.818817 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" containerName="init" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.818824 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" containerName="init" Nov 24 10:34:33 crc kubenswrapper[4944]: E1124 10:34:33.818866 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b22277-7b6e-457b-8dd1-4bd9bff48f80" containerName="init" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.818873 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b22277-7b6e-457b-8dd1-4bd9bff48f80" containerName="init" Nov 24 10:34:33 crc kubenswrapper[4944]: E1124 10:34:33.818884 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b22277-7b6e-457b-8dd1-4bd9bff48f80" containerName="dnsmasq-dns" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.818891 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b22277-7b6e-457b-8dd1-4bd9bff48f80" containerName="dnsmasq-dns" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.819208 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d773f7a-f45b-4bdd-8167-9e7415a7d7b6" containerName="dnsmasq-dns" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.819230 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b22277-7b6e-457b-8dd1-4bd9bff48f80" containerName="dnsmasq-dns" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.820187 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.827727 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.827960 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.828313 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.832497 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f"] Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.836561 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.868380 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.868454 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.868528 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j67gz\" (UniqueName: \"kubernetes.io/projected/35e9b789-d140-4c8e-af16-8a706a439e19-kube-api-access-j67gz\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.868567 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.868599 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.971945 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j67gz\" (UniqueName: \"kubernetes.io/projected/35e9b789-d140-4c8e-af16-8a706a439e19-kube-api-access-j67gz\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.972114 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.972186 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.972617 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.972689 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.979641 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.979936 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.981206 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.986998 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:33 crc kubenswrapper[4944]: I1124 10:34:33.992550 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j67gz\" (UniqueName: \"kubernetes.io/projected/35e9b789-d140-4c8e-af16-8a706a439e19-kube-api-access-j67gz\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:34 crc kubenswrapper[4944]: I1124 10:34:34.144884 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:34 crc kubenswrapper[4944]: I1124 10:34:34.216685 4944 scope.go:117] "RemoveContainer" containerID="c22915dd38e7aa912865e171d32307a63f691938a23c2eb7f333d8f873ff2379" Nov 24 10:34:34 crc kubenswrapper[4944]: I1124 10:34:34.332677 4944 scope.go:117] "RemoveContainer" containerID="b6f5c3afeedac08662dbac9c37ec2608bf5c9543bc0d6d9cabaa50a78e72513a" Nov 24 10:34:34 crc kubenswrapper[4944]: I1124 10:34:34.377442 4944 scope.go:117] "RemoveContainer" containerID="cd8c79a09e89d838bd4075b59169a50473fdf8071cd1897ceeaf005e16b8f143" Nov 24 10:34:34 crc kubenswrapper[4944]: W1124 10:34:34.771676 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35e9b789_d140_4c8e_af16_8a706a439e19.slice/crio-0b5a233167e43bbfd811015b5b37e55cb0296d99571f691edf54eaa9706551c6 WatchSource:0}: Error finding container 0b5a233167e43bbfd811015b5b37e55cb0296d99571f691edf54eaa9706551c6: Status 404 returned error can't find the container with id 0b5a233167e43bbfd811015b5b37e55cb0296d99571f691edf54eaa9706551c6 Nov 24 10:34:34 crc kubenswrapper[4944]: I1124 10:34:34.774970 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f"] Nov 24 10:34:35 crc kubenswrapper[4944]: I1124 10:34:35.065766 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" event={"ID":"35e9b789-d140-4c8e-af16-8a706a439e19","Type":"ContainerStarted","Data":"0b5a233167e43bbfd811015b5b37e55cb0296d99571f691edf54eaa9706551c6"} Nov 24 10:34:44 crc kubenswrapper[4944]: I1124 10:34:44.166598 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" event={"ID":"35e9b789-d140-4c8e-af16-8a706a439e19","Type":"ContainerStarted","Data":"302458ecb1ac3b417c124ed4777e3e2d245f0c8d097f2ec61a6387d17330dfa9"} Nov 24 10:34:44 crc kubenswrapper[4944]: I1124 10:34:44.195292 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" podStartSLOduration=2.715919721 podStartE2EDuration="11.195271474s" podCreationTimestamp="2025-11-24 10:34:33 +0000 UTC" firstStartedPulling="2025-11-24 10:34:34.775487859 +0000 UTC m=+6135.309928321" lastFinishedPulling="2025-11-24 10:34:43.254839612 +0000 UTC m=+6143.789280074" observedRunningTime="2025-11-24 10:34:44.19108218 +0000 UTC m=+6144.725522702" watchObservedRunningTime="2025-11-24 10:34:44.195271474 +0000 UTC m=+6144.729711936" Nov 24 10:34:53 crc kubenswrapper[4944]: I1124 10:34:53.548823 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:34:53 crc kubenswrapper[4944]: I1124 10:34:53.549457 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:34:56 crc kubenswrapper[4944]: I1124 10:34:56.334974 4944 generic.go:334] "Generic (PLEG): container finished" podID="35e9b789-d140-4c8e-af16-8a706a439e19" containerID="302458ecb1ac3b417c124ed4777e3e2d245f0c8d097f2ec61a6387d17330dfa9" exitCode=0 Nov 24 10:34:56 crc kubenswrapper[4944]: I1124 10:34:56.335037 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" event={"ID":"35e9b789-d140-4c8e-af16-8a706a439e19","Type":"ContainerDied","Data":"302458ecb1ac3b417c124ed4777e3e2d245f0c8d097f2ec61a6387d17330dfa9"} Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.154027 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.227021 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j67gz\" (UniqueName: \"kubernetes.io/projected/35e9b789-d140-4c8e-af16-8a706a439e19-kube-api-access-j67gz\") pod \"35e9b789-d140-4c8e-af16-8a706a439e19\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.227436 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-pre-adoption-validation-combined-ca-bundle\") pod \"35e9b789-d140-4c8e-af16-8a706a439e19\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.227512 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ceph\") pod \"35e9b789-d140-4c8e-af16-8a706a439e19\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.227629 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ssh-key\") pod \"35e9b789-d140-4c8e-af16-8a706a439e19\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.227740 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-inventory\") pod \"35e9b789-d140-4c8e-af16-8a706a439e19\" (UID: \"35e9b789-d140-4c8e-af16-8a706a439e19\") " Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.236751 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e9b789-d140-4c8e-af16-8a706a439e19-kube-api-access-j67gz" (OuterVolumeSpecName: "kube-api-access-j67gz") pod "35e9b789-d140-4c8e-af16-8a706a439e19" (UID: "35e9b789-d140-4c8e-af16-8a706a439e19"). InnerVolumeSpecName "kube-api-access-j67gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.236885 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "35e9b789-d140-4c8e-af16-8a706a439e19" (UID: "35e9b789-d140-4c8e-af16-8a706a439e19"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.247745 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ceph" (OuterVolumeSpecName: "ceph") pod "35e9b789-d140-4c8e-af16-8a706a439e19" (UID: "35e9b789-d140-4c8e-af16-8a706a439e19"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.265501 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-inventory" (OuterVolumeSpecName: "inventory") pod "35e9b789-d140-4c8e-af16-8a706a439e19" (UID: "35e9b789-d140-4c8e-af16-8a706a439e19"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.266795 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "35e9b789-d140-4c8e-af16-8a706a439e19" (UID: "35e9b789-d140-4c8e-af16-8a706a439e19"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.330386 4944 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.330415 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.330425 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.330536 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35e9b789-d140-4c8e-af16-8a706a439e19-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.330742 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j67gz\" (UniqueName: \"kubernetes.io/projected/35e9b789-d140-4c8e-af16-8a706a439e19-kube-api-access-j67gz\") on node \"crc\" DevicePath \"\"" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.360673 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" event={"ID":"35e9b789-d140-4c8e-af16-8a706a439e19","Type":"ContainerDied","Data":"0b5a233167e43bbfd811015b5b37e55cb0296d99571f691edf54eaa9706551c6"} Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.360713 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b5a233167e43bbfd811015b5b37e55cb0296d99571f691edf54eaa9706551c6" Nov 24 10:34:58 crc kubenswrapper[4944]: I1124 10:34:58.360720 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.415705 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs"] Nov 24 10:35:07 crc kubenswrapper[4944]: E1124 10:35:07.416681 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e9b789-d140-4c8e-af16-8a706a439e19" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.416695 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e9b789-d140-4c8e-af16-8a706a439e19" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.416897 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e9b789-d140-4c8e-af16-8a706a439e19" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.417722 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.425547 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.425997 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.426259 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.426393 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.430126 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs"] Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.535625 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.535787 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.535872 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.535926 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k8sd\" (UniqueName: \"kubernetes.io/projected/c6c0314e-cc78-4144-a276-5badbc29bb53-kube-api-access-8k8sd\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.536019 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.638105 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.638466 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k8sd\" (UniqueName: \"kubernetes.io/projected/c6c0314e-cc78-4144-a276-5badbc29bb53-kube-api-access-8k8sd\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.638605 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.638772 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.638983 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.644259 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.644638 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.644870 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.645000 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.657788 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k8sd\" (UniqueName: \"kubernetes.io/projected/c6c0314e-cc78-4144-a276-5badbc29bb53-kube-api-access-8k8sd\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:07 crc kubenswrapper[4944]: I1124 10:35:07.739391 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:35:08 crc kubenswrapper[4944]: I1124 10:35:08.303881 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs"] Nov 24 10:35:08 crc kubenswrapper[4944]: I1124 10:35:08.315980 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 10:35:08 crc kubenswrapper[4944]: I1124 10:35:08.472766 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" event={"ID":"c6c0314e-cc78-4144-a276-5badbc29bb53","Type":"ContainerStarted","Data":"dcb27202093be5a0f0ae3d300f39d60cb5ca283234d52a11e1f4c43e57b45ebd"} Nov 24 10:35:09 crc kubenswrapper[4944]: I1124 10:35:09.483561 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" event={"ID":"c6c0314e-cc78-4144-a276-5badbc29bb53","Type":"ContainerStarted","Data":"c0cfc766694bc4c9761699aad281f161f1f0da731f8dd248bf10f19b79988d71"} Nov 24 10:35:09 crc kubenswrapper[4944]: I1124 10:35:09.502847 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" podStartSLOduration=2.079317921 podStartE2EDuration="2.5028311s" podCreationTimestamp="2025-11-24 10:35:07 +0000 UTC" firstStartedPulling="2025-11-24 10:35:08.315411071 +0000 UTC m=+6168.849851533" lastFinishedPulling="2025-11-24 10:35:08.73892425 +0000 UTC m=+6169.273364712" observedRunningTime="2025-11-24 10:35:09.502266761 +0000 UTC m=+6170.036707233" watchObservedRunningTime="2025-11-24 10:35:09.5028311 +0000 UTC m=+6170.037271562" Nov 24 10:35:23 crc kubenswrapper[4944]: I1124 10:35:23.548041 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:35:23 crc kubenswrapper[4944]: I1124 10:35:23.548829 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:35:23 crc kubenswrapper[4944]: I1124 10:35:23.548873 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 10:35:23 crc kubenswrapper[4944]: I1124 10:35:23.549728 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c665bc43080f7d77b877e010e51181985aff09e212dee7ac1010f5e6186cdd1"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 10:35:23 crc kubenswrapper[4944]: I1124 10:35:23.549783 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://6c665bc43080f7d77b877e010e51181985aff09e212dee7ac1010f5e6186cdd1" gracePeriod=600 Nov 24 10:35:24 crc kubenswrapper[4944]: I1124 10:35:24.646763 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="6c665bc43080f7d77b877e010e51181985aff09e212dee7ac1010f5e6186cdd1" exitCode=0 Nov 24 10:35:24 crc kubenswrapper[4944]: I1124 10:35:24.646862 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"6c665bc43080f7d77b877e010e51181985aff09e212dee7ac1010f5e6186cdd1"} Nov 24 10:35:24 crc kubenswrapper[4944]: I1124 10:35:24.648285 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4"} Nov 24 10:35:24 crc kubenswrapper[4944]: I1124 10:35:24.648326 4944 scope.go:117] "RemoveContainer" containerID="90fbc468d27e5f8850dcb833c3220e37a51b53f798bde962670c1764e2ce3754" Nov 24 10:35:34 crc kubenswrapper[4944]: I1124 10:35:34.496692 4944 scope.go:117] "RemoveContainer" containerID="fe7e6a6f9a02cfcbcc024e48757dffc8f122d24577674139ea22a69099470c66" Nov 24 10:35:34 crc kubenswrapper[4944]: I1124 10:35:34.678601 4944 scope.go:117] "RemoveContainer" containerID="8069ddfd4b1bb0600b68b04547dcd3c9f426247aeaca6e14c31d7cef1f59305c" Nov 24 10:36:27 crc kubenswrapper[4944]: I1124 10:36:27.036509 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-ztssr"] Nov 24 10:36:27 crc kubenswrapper[4944]: I1124 10:36:27.046317 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-ztssr"] Nov 24 10:36:28 crc kubenswrapper[4944]: I1124 10:36:28.309101 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08a84af0-675d-48d4-a3e4-d6f42781af31" path="/var/lib/kubelet/pods/08a84af0-675d-48d4-a3e4-d6f42781af31/volumes" Nov 24 10:36:29 crc kubenswrapper[4944]: I1124 10:36:29.029889 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-9ac7-account-create-84fwr"] Nov 24 10:36:29 crc kubenswrapper[4944]: I1124 10:36:29.040195 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-9ac7-account-create-84fwr"] Nov 24 10:36:30 crc kubenswrapper[4944]: I1124 10:36:30.292147 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61d0cc15-19b3-4c86-aadc-2d1814094e32" path="/var/lib/kubelet/pods/61d0cc15-19b3-4c86-aadc-2d1814094e32/volumes" Nov 24 10:36:34 crc kubenswrapper[4944]: I1124 10:36:34.745509 4944 scope.go:117] "RemoveContainer" containerID="0b49c7f48cc2e8d3424d50a15800d94df3f6da14722928e9e85d00d6384c5a34" Nov 24 10:36:34 crc kubenswrapper[4944]: I1124 10:36:34.773344 4944 scope.go:117] "RemoveContainer" containerID="96e4207108cc6c18351e5deaefe376a84218b8730a50db7bbb94489337985754" Nov 24 10:36:35 crc kubenswrapper[4944]: I1124 10:36:35.039484 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-4hk8b"] Nov 24 10:36:35 crc kubenswrapper[4944]: I1124 10:36:35.055517 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-4hk8b"] Nov 24 10:36:36 crc kubenswrapper[4944]: I1124 10:36:36.036201 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-ed3e-account-create-cb28f"] Nov 24 10:36:36 crc kubenswrapper[4944]: I1124 10:36:36.048194 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-ed3e-account-create-cb28f"] Nov 24 10:36:36 crc kubenswrapper[4944]: I1124 10:36:36.290403 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2de39c88-cdcc-41db-9bdc-13b0d990a451" path="/var/lib/kubelet/pods/2de39c88-cdcc-41db-9bdc-13b0d990a451/volumes" Nov 24 10:36:36 crc kubenswrapper[4944]: I1124 10:36:36.292609 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7caf33c-af3c-4e3e-aba3-a7480fa69b85" path="/var/lib/kubelet/pods/b7caf33c-af3c-4e3e-aba3-a7480fa69b85/volumes" Nov 24 10:37:15 crc kubenswrapper[4944]: I1124 10:37:15.045189 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-hsl6x"] Nov 24 10:37:15 crc kubenswrapper[4944]: I1124 10:37:15.059226 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-hsl6x"] Nov 24 10:37:16 crc kubenswrapper[4944]: I1124 10:37:16.292635 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47d6bd13-b381-4117-84a3-5418a76f2d24" path="/var/lib/kubelet/pods/47d6bd13-b381-4117-84a3-5418a76f2d24/volumes" Nov 24 10:37:23 crc kubenswrapper[4944]: I1124 10:37:23.549270 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:37:23 crc kubenswrapper[4944]: I1124 10:37:23.550119 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:37:34 crc kubenswrapper[4944]: I1124 10:37:34.885822 4944 scope.go:117] "RemoveContainer" containerID="51daae17ad1be180b249965e9c850e8d7dbf2252bbbb4f1dbded9e2e35872525" Nov 24 10:37:34 crc kubenswrapper[4944]: I1124 10:37:34.920556 4944 scope.go:117] "RemoveContainer" containerID="0a5a72c6a6de28b80dace142d0f5de9523f8f98573e6ceb13cbf7eaaf2c48d82" Nov 24 10:37:34 crc kubenswrapper[4944]: I1124 10:37:34.983469 4944 scope.go:117] "RemoveContainer" containerID="110d451e530b0df396a8093a35d92ec7d64fd08c04ca4d482aaa619f0f3bf2ad" Nov 24 10:37:35 crc kubenswrapper[4944]: I1124 10:37:35.031942 4944 scope.go:117] "RemoveContainer" containerID="ab8146ee72b37ba2c492d32d88f38d907be18b127b614a85fb669e060992a9d7" Nov 24 10:37:53 crc kubenswrapper[4944]: I1124 10:37:53.548382 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:37:53 crc kubenswrapper[4944]: I1124 10:37:53.548967 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:38:23 crc kubenswrapper[4944]: I1124 10:38:23.548252 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:38:23 crc kubenswrapper[4944]: I1124 10:38:23.548825 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:38:23 crc kubenswrapper[4944]: I1124 10:38:23.548870 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 10:38:23 crc kubenswrapper[4944]: I1124 10:38:23.549505 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 10:38:23 crc kubenswrapper[4944]: I1124 10:38:23.549558 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" gracePeriod=600 Nov 24 10:38:23 crc kubenswrapper[4944]: E1124 10:38:23.669013 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:38:24 crc kubenswrapper[4944]: I1124 10:38:24.427714 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" exitCode=0 Nov 24 10:38:24 crc kubenswrapper[4944]: I1124 10:38:24.427781 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4"} Nov 24 10:38:24 crc kubenswrapper[4944]: I1124 10:38:24.428526 4944 scope.go:117] "RemoveContainer" containerID="6c665bc43080f7d77b877e010e51181985aff09e212dee7ac1010f5e6186cdd1" Nov 24 10:38:24 crc kubenswrapper[4944]: I1124 10:38:24.430720 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:38:24 crc kubenswrapper[4944]: E1124 10:38:24.431275 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:38:39 crc kubenswrapper[4944]: I1124 10:38:39.279394 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:38:39 crc kubenswrapper[4944]: E1124 10:38:39.281515 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:38:52 crc kubenswrapper[4944]: I1124 10:38:52.277874 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:38:52 crc kubenswrapper[4944]: E1124 10:38:52.278845 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:39:03 crc kubenswrapper[4944]: I1124 10:39:03.277920 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:39:03 crc kubenswrapper[4944]: E1124 10:39:03.278638 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:39:17 crc kubenswrapper[4944]: I1124 10:39:17.277373 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:39:17 crc kubenswrapper[4944]: E1124 10:39:17.278066 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:39:29 crc kubenswrapper[4944]: I1124 10:39:29.277120 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:39:29 crc kubenswrapper[4944]: E1124 10:39:29.277886 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:39:35 crc kubenswrapper[4944]: I1124 10:39:35.180143 4944 scope.go:117] "RemoveContainer" containerID="0d3f325481b0db6a0fb974381abfa1af0d973fc3fb9f9963c3327bcbe395aa7a" Nov 24 10:39:35 crc kubenswrapper[4944]: I1124 10:39:35.202846 4944 scope.go:117] "RemoveContainer" containerID="daf896dace00283836bfcd6ed0fa4131ebf3fc3cb1686c2de33706e0f59e50e4" Nov 24 10:39:40 crc kubenswrapper[4944]: I1124 10:39:40.284339 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:39:40 crc kubenswrapper[4944]: E1124 10:39:40.285151 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:39:44 crc kubenswrapper[4944]: I1124 10:39:44.057913 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-eaba-account-create-wdpst"] Nov 24 10:39:44 crc kubenswrapper[4944]: I1124 10:39:44.074421 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-9l7mt"] Nov 24 10:39:44 crc kubenswrapper[4944]: I1124 10:39:44.085724 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-eaba-account-create-wdpst"] Nov 24 10:39:44 crc kubenswrapper[4944]: I1124 10:39:44.095860 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-9l7mt"] Nov 24 10:39:44 crc kubenswrapper[4944]: I1124 10:39:44.287180 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95067a49-0792-4ee9-a0fd-6cec20b39955" path="/var/lib/kubelet/pods/95067a49-0792-4ee9-a0fd-6cec20b39955/volumes" Nov 24 10:39:44 crc kubenswrapper[4944]: I1124 10:39:44.288220 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbca7358-993f-4558-b9b3-e8f41abfbcdf" path="/var/lib/kubelet/pods/dbca7358-993f-4558-b9b3-e8f41abfbcdf/volumes" Nov 24 10:39:52 crc kubenswrapper[4944]: I1124 10:39:52.277564 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:39:52 crc kubenswrapper[4944]: E1124 10:39:52.278439 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:39:56 crc kubenswrapper[4944]: I1124 10:39:56.033713 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-szw7j"] Nov 24 10:39:56 crc kubenswrapper[4944]: I1124 10:39:56.042834 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-szw7j"] Nov 24 10:39:56 crc kubenswrapper[4944]: I1124 10:39:56.292773 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bc364f9-82a1-4f2f-9eed-980aa56e8e57" path="/var/lib/kubelet/pods/8bc364f9-82a1-4f2f-9eed-980aa56e8e57/volumes" Nov 24 10:40:04 crc kubenswrapper[4944]: I1124 10:40:04.297743 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:40:04 crc kubenswrapper[4944]: E1124 10:40:04.298644 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:40:15 crc kubenswrapper[4944]: I1124 10:40:15.277558 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:40:15 crc kubenswrapper[4944]: E1124 10:40:15.278419 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:40:28 crc kubenswrapper[4944]: I1124 10:40:28.277988 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:40:28 crc kubenswrapper[4944]: E1124 10:40:28.279235 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:40:35 crc kubenswrapper[4944]: I1124 10:40:35.257757 4944 scope.go:117] "RemoveContainer" containerID="10a3d01cc66611bf337fd55105a30d5361b82dbad1d1968259ab6a4c479f5c45" Nov 24 10:40:35 crc kubenswrapper[4944]: I1124 10:40:35.281361 4944 scope.go:117] "RemoveContainer" containerID="53f6c31fbb3ae4d8e49ccc5f3e84a584be9399669eaf919e56a751eb125e09ac" Nov 24 10:40:35 crc kubenswrapper[4944]: I1124 10:40:35.342556 4944 scope.go:117] "RemoveContainer" containerID="f4f7775af7d56e1d996b230ef9d8a67199ec7f3c69c55cc3f8ca797cc1fc289e" Nov 24 10:40:40 crc kubenswrapper[4944]: I1124 10:40:40.289123 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:40:40 crc kubenswrapper[4944]: E1124 10:40:40.289996 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:40:52 crc kubenswrapper[4944]: I1124 10:40:52.276979 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:40:52 crc kubenswrapper[4944]: E1124 10:40:52.277763 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:41:06 crc kubenswrapper[4944]: I1124 10:41:06.277637 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:41:06 crc kubenswrapper[4944]: E1124 10:41:06.278533 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:41:19 crc kubenswrapper[4944]: I1124 10:41:19.276825 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:41:19 crc kubenswrapper[4944]: E1124 10:41:19.277484 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:41:33 crc kubenswrapper[4944]: I1124 10:41:33.277332 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:41:33 crc kubenswrapper[4944]: E1124 10:41:33.278072 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:41:48 crc kubenswrapper[4944]: I1124 10:41:48.277131 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:41:48 crc kubenswrapper[4944]: E1124 10:41:48.277914 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:41:56 crc kubenswrapper[4944]: I1124 10:41:56.811438 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5cxpb"] Nov 24 10:41:56 crc kubenswrapper[4944]: I1124 10:41:56.820735 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:41:56 crc kubenswrapper[4944]: I1124 10:41:56.867225 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5cxpb"] Nov 24 10:41:56 crc kubenswrapper[4944]: I1124 10:41:56.927707 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vnqz\" (UniqueName: \"kubernetes.io/projected/471a253d-9cc9-43d2-8550-9bf6007bf49a-kube-api-access-2vnqz\") pod \"redhat-operators-5cxpb\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:41:56 crc kubenswrapper[4944]: I1124 10:41:56.927790 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-catalog-content\") pod \"redhat-operators-5cxpb\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:41:56 crc kubenswrapper[4944]: I1124 10:41:56.927858 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-utilities\") pod \"redhat-operators-5cxpb\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:41:57 crc kubenswrapper[4944]: I1124 10:41:57.030565 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vnqz\" (UniqueName: \"kubernetes.io/projected/471a253d-9cc9-43d2-8550-9bf6007bf49a-kube-api-access-2vnqz\") pod \"redhat-operators-5cxpb\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:41:57 crc kubenswrapper[4944]: I1124 10:41:57.030655 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-catalog-content\") pod \"redhat-operators-5cxpb\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:41:57 crc kubenswrapper[4944]: I1124 10:41:57.030713 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-utilities\") pod \"redhat-operators-5cxpb\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:41:57 crc kubenswrapper[4944]: I1124 10:41:57.031429 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-catalog-content\") pod \"redhat-operators-5cxpb\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:41:57 crc kubenswrapper[4944]: I1124 10:41:57.031464 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-utilities\") pod \"redhat-operators-5cxpb\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:41:57 crc kubenswrapper[4944]: I1124 10:41:57.060141 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vnqz\" (UniqueName: \"kubernetes.io/projected/471a253d-9cc9-43d2-8550-9bf6007bf49a-kube-api-access-2vnqz\") pod \"redhat-operators-5cxpb\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:41:57 crc kubenswrapper[4944]: I1124 10:41:57.150750 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:41:57 crc kubenswrapper[4944]: I1124 10:41:57.647563 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5cxpb"] Nov 24 10:41:58 crc kubenswrapper[4944]: I1124 10:41:58.662603 4944 generic.go:334] "Generic (PLEG): container finished" podID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerID="94c61fc459ff2fd4e68749187980c9dc43a7358c5bf60d1bc0f470fa0b03d1f7" exitCode=0 Nov 24 10:41:58 crc kubenswrapper[4944]: I1124 10:41:58.662836 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cxpb" event={"ID":"471a253d-9cc9-43d2-8550-9bf6007bf49a","Type":"ContainerDied","Data":"94c61fc459ff2fd4e68749187980c9dc43a7358c5bf60d1bc0f470fa0b03d1f7"} Nov 24 10:41:58 crc kubenswrapper[4944]: I1124 10:41:58.663135 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cxpb" event={"ID":"471a253d-9cc9-43d2-8550-9bf6007bf49a","Type":"ContainerStarted","Data":"77af3d920560e249bf2e40143d835661e329e7c17dfcbf863769b65e868d13bf"} Nov 24 10:41:58 crc kubenswrapper[4944]: I1124 10:41:58.665535 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 10:41:59 crc kubenswrapper[4944]: I1124 10:41:59.675424 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cxpb" event={"ID":"471a253d-9cc9-43d2-8550-9bf6007bf49a","Type":"ContainerStarted","Data":"3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd"} Nov 24 10:42:03 crc kubenswrapper[4944]: I1124 10:42:03.278830 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:42:03 crc kubenswrapper[4944]: E1124 10:42:03.279605 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:42:03 crc kubenswrapper[4944]: I1124 10:42:03.710283 4944 generic.go:334] "Generic (PLEG): container finished" podID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerID="3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd" exitCode=0 Nov 24 10:42:03 crc kubenswrapper[4944]: I1124 10:42:03.710352 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cxpb" event={"ID":"471a253d-9cc9-43d2-8550-9bf6007bf49a","Type":"ContainerDied","Data":"3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd"} Nov 24 10:42:04 crc kubenswrapper[4944]: I1124 10:42:04.721338 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cxpb" event={"ID":"471a253d-9cc9-43d2-8550-9bf6007bf49a","Type":"ContainerStarted","Data":"2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25"} Nov 24 10:42:04 crc kubenswrapper[4944]: I1124 10:42:04.736830 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5cxpb" podStartSLOduration=3.05391482 podStartE2EDuration="8.736813909s" podCreationTimestamp="2025-11-24 10:41:56 +0000 UTC" firstStartedPulling="2025-11-24 10:41:58.665329154 +0000 UTC m=+6579.199769616" lastFinishedPulling="2025-11-24 10:42:04.348228243 +0000 UTC m=+6584.882668705" observedRunningTime="2025-11-24 10:42:04.736140987 +0000 UTC m=+6585.270581469" watchObservedRunningTime="2025-11-24 10:42:04.736813909 +0000 UTC m=+6585.271254371" Nov 24 10:42:07 crc kubenswrapper[4944]: I1124 10:42:07.151415 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:42:07 crc kubenswrapper[4944]: I1124 10:42:07.151826 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:42:08 crc kubenswrapper[4944]: I1124 10:42:08.219646 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5cxpb" podUID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerName="registry-server" probeResult="failure" output=< Nov 24 10:42:08 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 10:42:08 crc kubenswrapper[4944]: > Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.197908 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dq5xk"] Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.200261 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.229145 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dq5xk"] Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.315837 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-utilities\") pod \"certified-operators-dq5xk\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.315890 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-catalog-content\") pod \"certified-operators-dq5xk\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.316134 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvfdp\" (UniqueName: \"kubernetes.io/projected/103883a0-b78c-42ac-99f3-15d606eeac23-kube-api-access-fvfdp\") pod \"certified-operators-dq5xk\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.417590 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-utilities\") pod \"certified-operators-dq5xk\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.417631 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-catalog-content\") pod \"certified-operators-dq5xk\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.417715 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvfdp\" (UniqueName: \"kubernetes.io/projected/103883a0-b78c-42ac-99f3-15d606eeac23-kube-api-access-fvfdp\") pod \"certified-operators-dq5xk\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.418106 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-utilities\") pod \"certified-operators-dq5xk\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.418107 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-catalog-content\") pod \"certified-operators-dq5xk\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.438239 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvfdp\" (UniqueName: \"kubernetes.io/projected/103883a0-b78c-42ac-99f3-15d606eeac23-kube-api-access-fvfdp\") pod \"certified-operators-dq5xk\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:09 crc kubenswrapper[4944]: I1124 10:42:09.549631 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:10 crc kubenswrapper[4944]: I1124 10:42:10.217718 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dq5xk"] Nov 24 10:42:10 crc kubenswrapper[4944]: I1124 10:42:10.777555 4944 generic.go:334] "Generic (PLEG): container finished" podID="103883a0-b78c-42ac-99f3-15d606eeac23" containerID="c10305958bc7dec1a983a82557ab7821f46cdf6da5da007a6ae59244dc7995cf" exitCode=0 Nov 24 10:42:10 crc kubenswrapper[4944]: I1124 10:42:10.777610 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq5xk" event={"ID":"103883a0-b78c-42ac-99f3-15d606eeac23","Type":"ContainerDied","Data":"c10305958bc7dec1a983a82557ab7821f46cdf6da5da007a6ae59244dc7995cf"} Nov 24 10:42:10 crc kubenswrapper[4944]: I1124 10:42:10.777854 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq5xk" event={"ID":"103883a0-b78c-42ac-99f3-15d606eeac23","Type":"ContainerStarted","Data":"1005343a45c9151b07f43768e97725539090deb5e0eef4144a826a9acb4e206e"} Nov 24 10:42:11 crc kubenswrapper[4944]: I1124 10:42:11.787981 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq5xk" event={"ID":"103883a0-b78c-42ac-99f3-15d606eeac23","Type":"ContainerStarted","Data":"e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e"} Nov 24 10:42:13 crc kubenswrapper[4944]: I1124 10:42:13.810145 4944 generic.go:334] "Generic (PLEG): container finished" podID="103883a0-b78c-42ac-99f3-15d606eeac23" containerID="e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e" exitCode=0 Nov 24 10:42:13 crc kubenswrapper[4944]: I1124 10:42:13.810238 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq5xk" event={"ID":"103883a0-b78c-42ac-99f3-15d606eeac23","Type":"ContainerDied","Data":"e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e"} Nov 24 10:42:14 crc kubenswrapper[4944]: I1124 10:42:14.043545 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-pkg8k"] Nov 24 10:42:14 crc kubenswrapper[4944]: I1124 10:42:14.054241 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-2753-account-create-q5c4z"] Nov 24 10:42:14 crc kubenswrapper[4944]: I1124 10:42:14.062787 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-pkg8k"] Nov 24 10:42:14 crc kubenswrapper[4944]: I1124 10:42:14.070911 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-2753-account-create-q5c4z"] Nov 24 10:42:14 crc kubenswrapper[4944]: I1124 10:42:14.278953 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:42:14 crc kubenswrapper[4944]: E1124 10:42:14.279367 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:42:14 crc kubenswrapper[4944]: I1124 10:42:14.294129 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a" path="/var/lib/kubelet/pods/8cac3c5a-0baa-4f86-ba76-6d86e8ccb41a/volumes" Nov 24 10:42:14 crc kubenswrapper[4944]: I1124 10:42:14.295532 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee0dc0ee-3410-4c15-84d4-a26d5b65405a" path="/var/lib/kubelet/pods/ee0dc0ee-3410-4c15-84d4-a26d5b65405a/volumes" Nov 24 10:42:14 crc kubenswrapper[4944]: I1124 10:42:14.820803 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq5xk" event={"ID":"103883a0-b78c-42ac-99f3-15d606eeac23","Type":"ContainerStarted","Data":"56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a"} Nov 24 10:42:14 crc kubenswrapper[4944]: I1124 10:42:14.851199 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dq5xk" podStartSLOduration=2.410444703 podStartE2EDuration="5.851179448s" podCreationTimestamp="2025-11-24 10:42:09 +0000 UTC" firstStartedPulling="2025-11-24 10:42:10.779102614 +0000 UTC m=+6591.313543076" lastFinishedPulling="2025-11-24 10:42:14.219837369 +0000 UTC m=+6594.754277821" observedRunningTime="2025-11-24 10:42:14.843772052 +0000 UTC m=+6595.378212514" watchObservedRunningTime="2025-11-24 10:42:14.851179448 +0000 UTC m=+6595.385619910" Nov 24 10:42:17 crc kubenswrapper[4944]: I1124 10:42:17.211966 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:42:17 crc kubenswrapper[4944]: I1124 10:42:17.268397 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:42:17 crc kubenswrapper[4944]: I1124 10:42:17.788928 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5cxpb"] Nov 24 10:42:18 crc kubenswrapper[4944]: I1124 10:42:18.855212 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5cxpb" podUID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerName="registry-server" containerID="cri-o://2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25" gracePeriod=2 Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.353139 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.530276 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vnqz\" (UniqueName: \"kubernetes.io/projected/471a253d-9cc9-43d2-8550-9bf6007bf49a-kube-api-access-2vnqz\") pod \"471a253d-9cc9-43d2-8550-9bf6007bf49a\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.530449 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-catalog-content\") pod \"471a253d-9cc9-43d2-8550-9bf6007bf49a\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.531007 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-utilities\") pod \"471a253d-9cc9-43d2-8550-9bf6007bf49a\" (UID: \"471a253d-9cc9-43d2-8550-9bf6007bf49a\") " Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.531693 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-utilities" (OuterVolumeSpecName: "utilities") pod "471a253d-9cc9-43d2-8550-9bf6007bf49a" (UID: "471a253d-9cc9-43d2-8550-9bf6007bf49a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.535944 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/471a253d-9cc9-43d2-8550-9bf6007bf49a-kube-api-access-2vnqz" (OuterVolumeSpecName: "kube-api-access-2vnqz") pod "471a253d-9cc9-43d2-8550-9bf6007bf49a" (UID: "471a253d-9cc9-43d2-8550-9bf6007bf49a"). InnerVolumeSpecName "kube-api-access-2vnqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.550128 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.550185 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.599912 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.632072 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "471a253d-9cc9-43d2-8550-9bf6007bf49a" (UID: "471a253d-9cc9-43d2-8550-9bf6007bf49a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.633601 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.633626 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vnqz\" (UniqueName: \"kubernetes.io/projected/471a253d-9cc9-43d2-8550-9bf6007bf49a-kube-api-access-2vnqz\") on node \"crc\" DevicePath \"\"" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.633636 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a253d-9cc9-43d2-8550-9bf6007bf49a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.867167 4944 generic.go:334] "Generic (PLEG): container finished" podID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerID="2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25" exitCode=0 Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.867230 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5cxpb" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.867260 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cxpb" event={"ID":"471a253d-9cc9-43d2-8550-9bf6007bf49a","Type":"ContainerDied","Data":"2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25"} Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.867644 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5cxpb" event={"ID":"471a253d-9cc9-43d2-8550-9bf6007bf49a","Type":"ContainerDied","Data":"77af3d920560e249bf2e40143d835661e329e7c17dfcbf863769b65e868d13bf"} Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.867675 4944 scope.go:117] "RemoveContainer" containerID="2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.894354 4944 scope.go:117] "RemoveContainer" containerID="3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.905563 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5cxpb"] Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.917342 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5cxpb"] Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.919103 4944 scope.go:117] "RemoveContainer" containerID="94c61fc459ff2fd4e68749187980c9dc43a7358c5bf60d1bc0f470fa0b03d1f7" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.937313 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.985166 4944 scope.go:117] "RemoveContainer" containerID="2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25" Nov 24 10:42:19 crc kubenswrapper[4944]: E1124 10:42:19.985554 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25\": container with ID starting with 2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25 not found: ID does not exist" containerID="2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.985585 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25"} err="failed to get container status \"2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25\": rpc error: code = NotFound desc = could not find container \"2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25\": container with ID starting with 2c30a3f5ecc51b209142ed751408a680369c3511acd251e32cd0e77e82e31b25 not found: ID does not exist" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.985605 4944 scope.go:117] "RemoveContainer" containerID="3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd" Nov 24 10:42:19 crc kubenswrapper[4944]: E1124 10:42:19.985839 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd\": container with ID starting with 3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd not found: ID does not exist" containerID="3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.985868 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd"} err="failed to get container status \"3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd\": rpc error: code = NotFound desc = could not find container \"3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd\": container with ID starting with 3c45e7fca248e366e688db200ec6e44913e1ddcdc3fb4fd0532038bd8e08acfd not found: ID does not exist" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.985882 4944 scope.go:117] "RemoveContainer" containerID="94c61fc459ff2fd4e68749187980c9dc43a7358c5bf60d1bc0f470fa0b03d1f7" Nov 24 10:42:19 crc kubenswrapper[4944]: E1124 10:42:19.986070 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94c61fc459ff2fd4e68749187980c9dc43a7358c5bf60d1bc0f470fa0b03d1f7\": container with ID starting with 94c61fc459ff2fd4e68749187980c9dc43a7358c5bf60d1bc0f470fa0b03d1f7 not found: ID does not exist" containerID="94c61fc459ff2fd4e68749187980c9dc43a7358c5bf60d1bc0f470fa0b03d1f7" Nov 24 10:42:19 crc kubenswrapper[4944]: I1124 10:42:19.986091 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94c61fc459ff2fd4e68749187980c9dc43a7358c5bf60d1bc0f470fa0b03d1f7"} err="failed to get container status \"94c61fc459ff2fd4e68749187980c9dc43a7358c5bf60d1bc0f470fa0b03d1f7\": rpc error: code = NotFound desc = could not find container \"94c61fc459ff2fd4e68749187980c9dc43a7358c5bf60d1bc0f470fa0b03d1f7\": container with ID starting with 94c61fc459ff2fd4e68749187980c9dc43a7358c5bf60d1bc0f470fa0b03d1f7 not found: ID does not exist" Nov 24 10:42:20 crc kubenswrapper[4944]: I1124 10:42:20.287057 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="471a253d-9cc9-43d2-8550-9bf6007bf49a" path="/var/lib/kubelet/pods/471a253d-9cc9-43d2-8550-9bf6007bf49a/volumes" Nov 24 10:42:21 crc kubenswrapper[4944]: I1124 10:42:21.991319 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dq5xk"] Nov 24 10:42:21 crc kubenswrapper[4944]: I1124 10:42:21.991852 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dq5xk" podUID="103883a0-b78c-42ac-99f3-15d606eeac23" containerName="registry-server" containerID="cri-o://56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a" gracePeriod=2 Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.490128 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.602194 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-utilities\") pod \"103883a0-b78c-42ac-99f3-15d606eeac23\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.602536 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-catalog-content\") pod \"103883a0-b78c-42ac-99f3-15d606eeac23\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.602604 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvfdp\" (UniqueName: \"kubernetes.io/projected/103883a0-b78c-42ac-99f3-15d606eeac23-kube-api-access-fvfdp\") pod \"103883a0-b78c-42ac-99f3-15d606eeac23\" (UID: \"103883a0-b78c-42ac-99f3-15d606eeac23\") " Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.605246 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-utilities" (OuterVolumeSpecName: "utilities") pod "103883a0-b78c-42ac-99f3-15d606eeac23" (UID: "103883a0-b78c-42ac-99f3-15d606eeac23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.612609 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/103883a0-b78c-42ac-99f3-15d606eeac23-kube-api-access-fvfdp" (OuterVolumeSpecName: "kube-api-access-fvfdp") pod "103883a0-b78c-42ac-99f3-15d606eeac23" (UID: "103883a0-b78c-42ac-99f3-15d606eeac23"). InnerVolumeSpecName "kube-api-access-fvfdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.658304 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "103883a0-b78c-42ac-99f3-15d606eeac23" (UID: "103883a0-b78c-42ac-99f3-15d606eeac23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.704699 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.704730 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvfdp\" (UniqueName: \"kubernetes.io/projected/103883a0-b78c-42ac-99f3-15d606eeac23-kube-api-access-fvfdp\") on node \"crc\" DevicePath \"\"" Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.704740 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/103883a0-b78c-42ac-99f3-15d606eeac23-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.900734 4944 generic.go:334] "Generic (PLEG): container finished" podID="103883a0-b78c-42ac-99f3-15d606eeac23" containerID="56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a" exitCode=0 Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.900776 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq5xk" event={"ID":"103883a0-b78c-42ac-99f3-15d606eeac23","Type":"ContainerDied","Data":"56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a"} Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.900806 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq5xk" event={"ID":"103883a0-b78c-42ac-99f3-15d606eeac23","Type":"ContainerDied","Data":"1005343a45c9151b07f43768e97725539090deb5e0eef4144a826a9acb4e206e"} Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.900823 4944 scope.go:117] "RemoveContainer" containerID="56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a" Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.900815 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dq5xk" Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.920362 4944 scope.go:117] "RemoveContainer" containerID="e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e" Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.942213 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dq5xk"] Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.950939 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dq5xk"] Nov 24 10:42:22 crc kubenswrapper[4944]: I1124 10:42:22.967639 4944 scope.go:117] "RemoveContainer" containerID="c10305958bc7dec1a983a82557ab7821f46cdf6da5da007a6ae59244dc7995cf" Nov 24 10:42:23 crc kubenswrapper[4944]: I1124 10:42:23.001348 4944 scope.go:117] "RemoveContainer" containerID="56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a" Nov 24 10:42:23 crc kubenswrapper[4944]: E1124 10:42:23.001774 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a\": container with ID starting with 56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a not found: ID does not exist" containerID="56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a" Nov 24 10:42:23 crc kubenswrapper[4944]: I1124 10:42:23.001807 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a"} err="failed to get container status \"56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a\": rpc error: code = NotFound desc = could not find container \"56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a\": container with ID starting with 56789e5f5793a9490ddaf4dc63b6bd8e5de2a9a01fdc11460b6c908a011c040a not found: ID does not exist" Nov 24 10:42:23 crc kubenswrapper[4944]: I1124 10:42:23.001829 4944 scope.go:117] "RemoveContainer" containerID="e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e" Nov 24 10:42:23 crc kubenswrapper[4944]: E1124 10:42:23.002153 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e\": container with ID starting with e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e not found: ID does not exist" containerID="e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e" Nov 24 10:42:23 crc kubenswrapper[4944]: I1124 10:42:23.002192 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e"} err="failed to get container status \"e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e\": rpc error: code = NotFound desc = could not find container \"e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e\": container with ID starting with e4d942d44757d5b0b2ab62be13f0e91e479904eff5ad621ec9b334c2337f9d0e not found: ID does not exist" Nov 24 10:42:23 crc kubenswrapper[4944]: I1124 10:42:23.002218 4944 scope.go:117] "RemoveContainer" containerID="c10305958bc7dec1a983a82557ab7821f46cdf6da5da007a6ae59244dc7995cf" Nov 24 10:42:23 crc kubenswrapper[4944]: E1124 10:42:23.002497 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c10305958bc7dec1a983a82557ab7821f46cdf6da5da007a6ae59244dc7995cf\": container with ID starting with c10305958bc7dec1a983a82557ab7821f46cdf6da5da007a6ae59244dc7995cf not found: ID does not exist" containerID="c10305958bc7dec1a983a82557ab7821f46cdf6da5da007a6ae59244dc7995cf" Nov 24 10:42:23 crc kubenswrapper[4944]: I1124 10:42:23.002531 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c10305958bc7dec1a983a82557ab7821f46cdf6da5da007a6ae59244dc7995cf"} err="failed to get container status \"c10305958bc7dec1a983a82557ab7821f46cdf6da5da007a6ae59244dc7995cf\": rpc error: code = NotFound desc = could not find container \"c10305958bc7dec1a983a82557ab7821f46cdf6da5da007a6ae59244dc7995cf\": container with ID starting with c10305958bc7dec1a983a82557ab7821f46cdf6da5da007a6ae59244dc7995cf not found: ID does not exist" Nov 24 10:42:24 crc kubenswrapper[4944]: I1124 10:42:24.288892 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="103883a0-b78c-42ac-99f3-15d606eeac23" path="/var/lib/kubelet/pods/103883a0-b78c-42ac-99f3-15d606eeac23/volumes" Nov 24 10:42:25 crc kubenswrapper[4944]: I1124 10:42:25.031215 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-jr84l"] Nov 24 10:42:25 crc kubenswrapper[4944]: I1124 10:42:25.046122 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-jr84l"] Nov 24 10:42:26 crc kubenswrapper[4944]: I1124 10:42:26.290194 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eda0f187-a1ab-41f0-a90b-5123d728621e" path="/var/lib/kubelet/pods/eda0f187-a1ab-41f0-a90b-5123d728621e/volumes" Nov 24 10:42:29 crc kubenswrapper[4944]: I1124 10:42:29.277528 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:42:29 crc kubenswrapper[4944]: E1124 10:42:29.278696 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:42:35 crc kubenswrapper[4944]: I1124 10:42:35.496535 4944 scope.go:117] "RemoveContainer" containerID="1bdefd3990b35b37499638c60ea1d6197ce93dc301745500112073b1189459fe" Nov 24 10:42:35 crc kubenswrapper[4944]: I1124 10:42:35.518105 4944 scope.go:117] "RemoveContainer" containerID="5ad2a97afbfece7ee4ba94d839dd6698a890f1ee187b05601ef833f71891e1ea" Nov 24 10:42:35 crc kubenswrapper[4944]: I1124 10:42:35.575858 4944 scope.go:117] "RemoveContainer" containerID="84ee30ef46f7d24d23507c81cd6777c1b4457a8f974adf970ef9128c7420b090" Nov 24 10:42:41 crc kubenswrapper[4944]: I1124 10:42:41.040439 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-t2nc4"] Nov 24 10:42:41 crc kubenswrapper[4944]: I1124 10:42:41.049806 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-t2nc4"] Nov 24 10:42:41 crc kubenswrapper[4944]: I1124 10:42:41.277258 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:42:41 crc kubenswrapper[4944]: E1124 10:42:41.277502 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:42:42 crc kubenswrapper[4944]: I1124 10:42:42.026480 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-9d4b-account-create-8pw4q"] Nov 24 10:42:42 crc kubenswrapper[4944]: I1124 10:42:42.035023 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-9d4b-account-create-8pw4q"] Nov 24 10:42:42 crc kubenswrapper[4944]: I1124 10:42:42.295866 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e1d1afa-085a-4cc1-b772-9880bb17630e" path="/var/lib/kubelet/pods/0e1d1afa-085a-4cc1-b772-9880bb17630e/volumes" Nov 24 10:42:42 crc kubenswrapper[4944]: I1124 10:42:42.296721 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c88c41d9-7b3e-45e9-8a93-a6efeb696c1a" path="/var/lib/kubelet/pods/c88c41d9-7b3e-45e9-8a93-a6efeb696c1a/volumes" Nov 24 10:42:52 crc kubenswrapper[4944]: I1124 10:42:52.277911 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:42:52 crc kubenswrapper[4944]: E1124 10:42:52.278971 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:42:53 crc kubenswrapper[4944]: I1124 10:42:53.041372 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-9b42b"] Nov 24 10:42:53 crc kubenswrapper[4944]: I1124 10:42:53.055367 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-9b42b"] Nov 24 10:42:54 crc kubenswrapper[4944]: I1124 10:42:54.288274 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="761360e0-d8c9-4a45-96a3-c90e19f721fc" path="/var/lib/kubelet/pods/761360e0-d8c9-4a45-96a3-c90e19f721fc/volumes" Nov 24 10:43:03 crc kubenswrapper[4944]: I1124 10:43:03.276777 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:43:03 crc kubenswrapper[4944]: E1124 10:43:03.277575 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:43:14 crc kubenswrapper[4944]: I1124 10:43:14.277000 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:43:14 crc kubenswrapper[4944]: E1124 10:43:14.277841 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:43:27 crc kubenswrapper[4944]: I1124 10:43:27.276425 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:43:27 crc kubenswrapper[4944]: I1124 10:43:27.620643 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"7a707b8a170f6fd3b0b500f93173c046a7d3db862dd0277d05c759ca4d9c9eb4"} Nov 24 10:43:35 crc kubenswrapper[4944]: I1124 10:43:35.705624 4944 scope.go:117] "RemoveContainer" containerID="b819bb57042c4fa74f8cc6d42e58401e0eab3372a35f0a778fc7f3ea44dfd42e" Nov 24 10:43:35 crc kubenswrapper[4944]: I1124 10:43:35.741283 4944 scope.go:117] "RemoveContainer" containerID="4fafff5ddb63d7ff80bc2a407fe14081b6cf4503524602e3393ea68dc9b2ab9d" Nov 24 10:43:35 crc kubenswrapper[4944]: I1124 10:43:35.803135 4944 scope.go:117] "RemoveContainer" containerID="5a38ae420edb5d72efdd021fb55317d671c8b93642d008d2a5b43b45bca5bfb7" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.550537 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ft4rh"] Nov 24 10:44:03 crc kubenswrapper[4944]: E1124 10:44:03.551432 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="103883a0-b78c-42ac-99f3-15d606eeac23" containerName="extract-utilities" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.551445 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="103883a0-b78c-42ac-99f3-15d606eeac23" containerName="extract-utilities" Nov 24 10:44:03 crc kubenswrapper[4944]: E1124 10:44:03.551477 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="103883a0-b78c-42ac-99f3-15d606eeac23" containerName="extract-content" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.551483 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="103883a0-b78c-42ac-99f3-15d606eeac23" containerName="extract-content" Nov 24 10:44:03 crc kubenswrapper[4944]: E1124 10:44:03.551493 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerName="extract-utilities" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.551499 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerName="extract-utilities" Nov 24 10:44:03 crc kubenswrapper[4944]: E1124 10:44:03.551521 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerName="extract-content" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.551527 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerName="extract-content" Nov 24 10:44:03 crc kubenswrapper[4944]: E1124 10:44:03.551538 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="103883a0-b78c-42ac-99f3-15d606eeac23" containerName="registry-server" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.551544 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="103883a0-b78c-42ac-99f3-15d606eeac23" containerName="registry-server" Nov 24 10:44:03 crc kubenswrapper[4944]: E1124 10:44:03.551559 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerName="registry-server" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.551565 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerName="registry-server" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.551747 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="471a253d-9cc9-43d2-8550-9bf6007bf49a" containerName="registry-server" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.551775 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="103883a0-b78c-42ac-99f3-15d606eeac23" containerName="registry-server" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.553250 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.569785 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ft4rh"] Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.669888 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-utilities\") pod \"community-operators-ft4rh\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.669980 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-catalog-content\") pod \"community-operators-ft4rh\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.670146 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnbt6\" (UniqueName: \"kubernetes.io/projected/17d10093-15b1-4cc2-869a-3ce6f2000234-kube-api-access-lnbt6\") pod \"community-operators-ft4rh\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.771958 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-utilities\") pod \"community-operators-ft4rh\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.772124 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-catalog-content\") pod \"community-operators-ft4rh\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.772146 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnbt6\" (UniqueName: \"kubernetes.io/projected/17d10093-15b1-4cc2-869a-3ce6f2000234-kube-api-access-lnbt6\") pod \"community-operators-ft4rh\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.772626 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-utilities\") pod \"community-operators-ft4rh\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.772637 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-catalog-content\") pod \"community-operators-ft4rh\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.794812 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnbt6\" (UniqueName: \"kubernetes.io/projected/17d10093-15b1-4cc2-869a-3ce6f2000234-kube-api-access-lnbt6\") pod \"community-operators-ft4rh\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:03 crc kubenswrapper[4944]: I1124 10:44:03.874447 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:04 crc kubenswrapper[4944]: I1124 10:44:04.391341 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ft4rh"] Nov 24 10:44:04 crc kubenswrapper[4944]: I1124 10:44:04.964341 4944 generic.go:334] "Generic (PLEG): container finished" podID="17d10093-15b1-4cc2-869a-3ce6f2000234" containerID="a0a85bcf2ffbdfaa5096c7f4cc1b08e8d4cbeed659ba4c0893d77fc0e18f46aa" exitCode=0 Nov 24 10:44:04 crc kubenswrapper[4944]: I1124 10:44:04.964395 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ft4rh" event={"ID":"17d10093-15b1-4cc2-869a-3ce6f2000234","Type":"ContainerDied","Data":"a0a85bcf2ffbdfaa5096c7f4cc1b08e8d4cbeed659ba4c0893d77fc0e18f46aa"} Nov 24 10:44:04 crc kubenswrapper[4944]: I1124 10:44:04.964423 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ft4rh" event={"ID":"17d10093-15b1-4cc2-869a-3ce6f2000234","Type":"ContainerStarted","Data":"d84f6de458325091741c27301a287958c54c96ce9aa4459684c07b4eb6da1a02"} Nov 24 10:44:05 crc kubenswrapper[4944]: I1124 10:44:05.975213 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ft4rh" event={"ID":"17d10093-15b1-4cc2-869a-3ce6f2000234","Type":"ContainerStarted","Data":"ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021"} Nov 24 10:44:06 crc kubenswrapper[4944]: I1124 10:44:06.987319 4944 generic.go:334] "Generic (PLEG): container finished" podID="17d10093-15b1-4cc2-869a-3ce6f2000234" containerID="ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021" exitCode=0 Nov 24 10:44:06 crc kubenswrapper[4944]: I1124 10:44:06.987476 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ft4rh" event={"ID":"17d10093-15b1-4cc2-869a-3ce6f2000234","Type":"ContainerDied","Data":"ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021"} Nov 24 10:44:07 crc kubenswrapper[4944]: I1124 10:44:07.998667 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ft4rh" event={"ID":"17d10093-15b1-4cc2-869a-3ce6f2000234","Type":"ContainerStarted","Data":"1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5"} Nov 24 10:44:08 crc kubenswrapper[4944]: I1124 10:44:08.033387 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ft4rh" podStartSLOduration=2.539952124 podStartE2EDuration="5.033366268s" podCreationTimestamp="2025-11-24 10:44:03 +0000 UTC" firstStartedPulling="2025-11-24 10:44:04.966872822 +0000 UTC m=+6705.501313284" lastFinishedPulling="2025-11-24 10:44:07.460286966 +0000 UTC m=+6707.994727428" observedRunningTime="2025-11-24 10:44:08.027731669 +0000 UTC m=+6708.562172131" watchObservedRunningTime="2025-11-24 10:44:08.033366268 +0000 UTC m=+6708.567806750" Nov 24 10:44:13 crc kubenswrapper[4944]: I1124 10:44:13.874808 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:13 crc kubenswrapper[4944]: I1124 10:44:13.875428 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:13 crc kubenswrapper[4944]: I1124 10:44:13.952275 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:14 crc kubenswrapper[4944]: I1124 10:44:14.142590 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:14 crc kubenswrapper[4944]: I1124 10:44:14.193228 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ft4rh"] Nov 24 10:44:16 crc kubenswrapper[4944]: I1124 10:44:16.085618 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ft4rh" podUID="17d10093-15b1-4cc2-869a-3ce6f2000234" containerName="registry-server" containerID="cri-o://1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5" gracePeriod=2 Nov 24 10:44:16 crc kubenswrapper[4944]: I1124 10:44:16.586693 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:16 crc kubenswrapper[4944]: I1124 10:44:16.694390 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-utilities\") pod \"17d10093-15b1-4cc2-869a-3ce6f2000234\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " Nov 24 10:44:16 crc kubenswrapper[4944]: I1124 10:44:16.694683 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-catalog-content\") pod \"17d10093-15b1-4cc2-869a-3ce6f2000234\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " Nov 24 10:44:16 crc kubenswrapper[4944]: I1124 10:44:16.694713 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnbt6\" (UniqueName: \"kubernetes.io/projected/17d10093-15b1-4cc2-869a-3ce6f2000234-kube-api-access-lnbt6\") pod \"17d10093-15b1-4cc2-869a-3ce6f2000234\" (UID: \"17d10093-15b1-4cc2-869a-3ce6f2000234\") " Nov 24 10:44:16 crc kubenswrapper[4944]: I1124 10:44:16.695328 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-utilities" (OuterVolumeSpecName: "utilities") pod "17d10093-15b1-4cc2-869a-3ce6f2000234" (UID: "17d10093-15b1-4cc2-869a-3ce6f2000234"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:44:16 crc kubenswrapper[4944]: I1124 10:44:16.695521 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:44:16 crc kubenswrapper[4944]: I1124 10:44:16.700002 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17d10093-15b1-4cc2-869a-3ce6f2000234-kube-api-access-lnbt6" (OuterVolumeSpecName: "kube-api-access-lnbt6") pod "17d10093-15b1-4cc2-869a-3ce6f2000234" (UID: "17d10093-15b1-4cc2-869a-3ce6f2000234"). InnerVolumeSpecName "kube-api-access-lnbt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:44:16 crc kubenswrapper[4944]: I1124 10:44:16.742659 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17d10093-15b1-4cc2-869a-3ce6f2000234" (UID: "17d10093-15b1-4cc2-869a-3ce6f2000234"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:44:16 crc kubenswrapper[4944]: I1124 10:44:16.797703 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17d10093-15b1-4cc2-869a-3ce6f2000234-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:44:16 crc kubenswrapper[4944]: I1124 10:44:16.797739 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnbt6\" (UniqueName: \"kubernetes.io/projected/17d10093-15b1-4cc2-869a-3ce6f2000234-kube-api-access-lnbt6\") on node \"crc\" DevicePath \"\"" Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.095577 4944 generic.go:334] "Generic (PLEG): container finished" podID="17d10093-15b1-4cc2-869a-3ce6f2000234" containerID="1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5" exitCode=0 Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.095622 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ft4rh" Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.095636 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ft4rh" event={"ID":"17d10093-15b1-4cc2-869a-3ce6f2000234","Type":"ContainerDied","Data":"1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5"} Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.096896 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ft4rh" event={"ID":"17d10093-15b1-4cc2-869a-3ce6f2000234","Type":"ContainerDied","Data":"d84f6de458325091741c27301a287958c54c96ce9aa4459684c07b4eb6da1a02"} Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.096918 4944 scope.go:117] "RemoveContainer" containerID="1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5" Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.129410 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ft4rh"] Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.129853 4944 scope.go:117] "RemoveContainer" containerID="ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021" Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.140575 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ft4rh"] Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.156828 4944 scope.go:117] "RemoveContainer" containerID="a0a85bcf2ffbdfaa5096c7f4cc1b08e8d4cbeed659ba4c0893d77fc0e18f46aa" Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.206126 4944 scope.go:117] "RemoveContainer" containerID="1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5" Nov 24 10:44:17 crc kubenswrapper[4944]: E1124 10:44:17.206768 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5\": container with ID starting with 1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5 not found: ID does not exist" containerID="1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5" Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.206799 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5"} err="failed to get container status \"1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5\": rpc error: code = NotFound desc = could not find container \"1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5\": container with ID starting with 1284af8d429e34935995eea3cf7080cedbf99d68893e62bf8a790f8e3d3f4da5 not found: ID does not exist" Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.206821 4944 scope.go:117] "RemoveContainer" containerID="ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021" Nov 24 10:44:17 crc kubenswrapper[4944]: E1124 10:44:17.207243 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021\": container with ID starting with ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021 not found: ID does not exist" containerID="ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021" Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.207264 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021"} err="failed to get container status \"ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021\": rpc error: code = NotFound desc = could not find container \"ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021\": container with ID starting with ebf31b4d1ad21455432954bfa72767eb09904f2cda1eaaa6f020bc737aac2021 not found: ID does not exist" Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.207277 4944 scope.go:117] "RemoveContainer" containerID="a0a85bcf2ffbdfaa5096c7f4cc1b08e8d4cbeed659ba4c0893d77fc0e18f46aa" Nov 24 10:44:17 crc kubenswrapper[4944]: E1124 10:44:17.207566 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0a85bcf2ffbdfaa5096c7f4cc1b08e8d4cbeed659ba4c0893d77fc0e18f46aa\": container with ID starting with a0a85bcf2ffbdfaa5096c7f4cc1b08e8d4cbeed659ba4c0893d77fc0e18f46aa not found: ID does not exist" containerID="a0a85bcf2ffbdfaa5096c7f4cc1b08e8d4cbeed659ba4c0893d77fc0e18f46aa" Nov 24 10:44:17 crc kubenswrapper[4944]: I1124 10:44:17.207586 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0a85bcf2ffbdfaa5096c7f4cc1b08e8d4cbeed659ba4c0893d77fc0e18f46aa"} err="failed to get container status \"a0a85bcf2ffbdfaa5096c7f4cc1b08e8d4cbeed659ba4c0893d77fc0e18f46aa\": rpc error: code = NotFound desc = could not find container \"a0a85bcf2ffbdfaa5096c7f4cc1b08e8d4cbeed659ba4c0893d77fc0e18f46aa\": container with ID starting with a0a85bcf2ffbdfaa5096c7f4cc1b08e8d4cbeed659ba4c0893d77fc0e18f46aa not found: ID does not exist" Nov 24 10:44:18 crc kubenswrapper[4944]: I1124 10:44:18.294120 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17d10093-15b1-4cc2-869a-3ce6f2000234" path="/var/lib/kubelet/pods/17d10093-15b1-4cc2-869a-3ce6f2000234/volumes" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.175403 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5"] Nov 24 10:45:00 crc kubenswrapper[4944]: E1124 10:45:00.176468 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d10093-15b1-4cc2-869a-3ce6f2000234" containerName="extract-utilities" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.176488 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d10093-15b1-4cc2-869a-3ce6f2000234" containerName="extract-utilities" Nov 24 10:45:00 crc kubenswrapper[4944]: E1124 10:45:00.176534 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d10093-15b1-4cc2-869a-3ce6f2000234" containerName="registry-server" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.176541 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d10093-15b1-4cc2-869a-3ce6f2000234" containerName="registry-server" Nov 24 10:45:00 crc kubenswrapper[4944]: E1124 10:45:00.176575 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17d10093-15b1-4cc2-869a-3ce6f2000234" containerName="extract-content" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.176581 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="17d10093-15b1-4cc2-869a-3ce6f2000234" containerName="extract-content" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.176788 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="17d10093-15b1-4cc2-869a-3ce6f2000234" containerName="registry-server" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.177677 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.181036 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.181478 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.187299 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5"] Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.301504 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-config-volume\") pod \"collect-profiles-29399685-8djd5\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.301584 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9fz4\" (UniqueName: \"kubernetes.io/projected/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-kube-api-access-n9fz4\") pod \"collect-profiles-29399685-8djd5\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.301968 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-secret-volume\") pod \"collect-profiles-29399685-8djd5\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.404019 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-config-volume\") pod \"collect-profiles-29399685-8djd5\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.404106 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9fz4\" (UniqueName: \"kubernetes.io/projected/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-kube-api-access-n9fz4\") pod \"collect-profiles-29399685-8djd5\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.404252 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-secret-volume\") pod \"collect-profiles-29399685-8djd5\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.405469 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-config-volume\") pod \"collect-profiles-29399685-8djd5\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.410432 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-secret-volume\") pod \"collect-profiles-29399685-8djd5\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.425021 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9fz4\" (UniqueName: \"kubernetes.io/projected/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-kube-api-access-n9fz4\") pod \"collect-profiles-29399685-8djd5\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.512038 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:00 crc kubenswrapper[4944]: I1124 10:45:00.957548 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5"] Nov 24 10:45:01 crc kubenswrapper[4944]: I1124 10:45:01.540852 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" event={"ID":"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0","Type":"ContainerStarted","Data":"90075532c080473cc712253ebd5c712e14bc3c81827c9a1b645898014a12a778"} Nov 24 10:45:01 crc kubenswrapper[4944]: I1124 10:45:01.541269 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" event={"ID":"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0","Type":"ContainerStarted","Data":"a26f0588f89b93db078a2d9b57e14b6fb20b5dd51774934a4cf6cb0aeba39cbe"} Nov 24 10:45:02 crc kubenswrapper[4944]: I1124 10:45:02.561603 4944 generic.go:334] "Generic (PLEG): container finished" podID="b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0" containerID="90075532c080473cc712253ebd5c712e14bc3c81827c9a1b645898014a12a778" exitCode=0 Nov 24 10:45:02 crc kubenswrapper[4944]: I1124 10:45:02.561679 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" event={"ID":"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0","Type":"ContainerDied","Data":"90075532c080473cc712253ebd5c712e14bc3c81827c9a1b645898014a12a778"} Nov 24 10:45:03 crc kubenswrapper[4944]: I1124 10:45:03.954989 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.081497 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-secret-volume\") pod \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.081560 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-config-volume\") pod \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.081776 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9fz4\" (UniqueName: \"kubernetes.io/projected/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-kube-api-access-n9fz4\") pod \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\" (UID: \"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0\") " Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.082780 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-config-volume" (OuterVolumeSpecName: "config-volume") pod "b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0" (UID: "b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.083106 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.087467 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0" (UID: "b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.089354 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-kube-api-access-n9fz4" (OuterVolumeSpecName: "kube-api-access-n9fz4") pod "b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0" (UID: "b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0"). InnerVolumeSpecName "kube-api-access-n9fz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.185114 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9fz4\" (UniqueName: \"kubernetes.io/projected/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-kube-api-access-n9fz4\") on node \"crc\" DevicePath \"\"" Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.185157 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.589009 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" event={"ID":"b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0","Type":"ContainerDied","Data":"a26f0588f89b93db078a2d9b57e14b6fb20b5dd51774934a4cf6cb0aeba39cbe"} Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.589085 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a26f0588f89b93db078a2d9b57e14b6fb20b5dd51774934a4cf6cb0aeba39cbe" Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.589033 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5" Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.627258 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4"] Nov 24 10:45:04 crc kubenswrapper[4944]: I1124 10:45:04.634680 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399640-kxqg4"] Nov 24 10:45:06 crc kubenswrapper[4944]: I1124 10:45:06.292024 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3afa46ba-4844-49ee-90c1-fc2f85ce40d3" path="/var/lib/kubelet/pods/3afa46ba-4844-49ee-90c1-fc2f85ce40d3/volumes" Nov 24 10:45:34 crc kubenswrapper[4944]: I1124 10:45:34.859488 4944 generic.go:334] "Generic (PLEG): container finished" podID="c6c0314e-cc78-4144-a276-5badbc29bb53" containerID="c0cfc766694bc4c9761699aad281f161f1f0da731f8dd248bf10f19b79988d71" exitCode=0 Nov 24 10:45:34 crc kubenswrapper[4944]: I1124 10:45:34.859566 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" event={"ID":"c6c0314e-cc78-4144-a276-5badbc29bb53","Type":"ContainerDied","Data":"c0cfc766694bc4c9761699aad281f161f1f0da731f8dd248bf10f19b79988d71"} Nov 24 10:45:35 crc kubenswrapper[4944]: I1124 10:45:35.972409 4944 scope.go:117] "RemoveContainer" containerID="e501dd114bb8d78c9c67c389a00602d5f3d274931da0360899d90d8c8febe40b" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.333721 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.468405 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-inventory\") pod \"c6c0314e-cc78-4144-a276-5badbc29bb53\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.468814 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-tripleo-cleanup-combined-ca-bundle\") pod \"c6c0314e-cc78-4144-a276-5badbc29bb53\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.468846 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k8sd\" (UniqueName: \"kubernetes.io/projected/c6c0314e-cc78-4144-a276-5badbc29bb53-kube-api-access-8k8sd\") pod \"c6c0314e-cc78-4144-a276-5badbc29bb53\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.468865 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ssh-key\") pod \"c6c0314e-cc78-4144-a276-5badbc29bb53\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.468916 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ceph\") pod \"c6c0314e-cc78-4144-a276-5badbc29bb53\" (UID: \"c6c0314e-cc78-4144-a276-5badbc29bb53\") " Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.475152 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "c6c0314e-cc78-4144-a276-5badbc29bb53" (UID: "c6c0314e-cc78-4144-a276-5badbc29bb53"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.475702 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6c0314e-cc78-4144-a276-5badbc29bb53-kube-api-access-8k8sd" (OuterVolumeSpecName: "kube-api-access-8k8sd") pod "c6c0314e-cc78-4144-a276-5badbc29bb53" (UID: "c6c0314e-cc78-4144-a276-5badbc29bb53"). InnerVolumeSpecName "kube-api-access-8k8sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.475790 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ceph" (OuterVolumeSpecName: "ceph") pod "c6c0314e-cc78-4144-a276-5badbc29bb53" (UID: "c6c0314e-cc78-4144-a276-5badbc29bb53"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.502126 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-inventory" (OuterVolumeSpecName: "inventory") pod "c6c0314e-cc78-4144-a276-5badbc29bb53" (UID: "c6c0314e-cc78-4144-a276-5badbc29bb53"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.504575 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c6c0314e-cc78-4144-a276-5badbc29bb53" (UID: "c6c0314e-cc78-4144-a276-5badbc29bb53"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.571806 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.571841 4944 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.571855 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k8sd\" (UniqueName: \"kubernetes.io/projected/c6c0314e-cc78-4144-a276-5badbc29bb53-kube-api-access-8k8sd\") on node \"crc\" DevicePath \"\"" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.571866 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.571876 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c6c0314e-cc78-4144-a276-5badbc29bb53-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.879355 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" event={"ID":"c6c0314e-cc78-4144-a276-5badbc29bb53","Type":"ContainerDied","Data":"dcb27202093be5a0f0ae3d300f39d60cb5ca283234d52a11e1f4c43e57b45ebd"} Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.879405 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcb27202093be5a0f0ae3d300f39d60cb5ca283234d52a11e1f4c43e57b45ebd" Nov 24 10:45:36 crc kubenswrapper[4944]: I1124 10:45:36.879465 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.120222 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-9pwb2"] Nov 24 10:45:47 crc kubenswrapper[4944]: E1124 10:45:47.122688 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0" containerName="collect-profiles" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.122802 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0" containerName="collect-profiles" Nov 24 10:45:47 crc kubenswrapper[4944]: E1124 10:45:47.122941 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6c0314e-cc78-4144-a276-5badbc29bb53" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.123617 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6c0314e-cc78-4144-a276-5badbc29bb53" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.123981 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0" containerName="collect-profiles" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.124121 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6c0314e-cc78-4144-a276-5badbc29bb53" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.125272 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.127778 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.128447 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.128447 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.128747 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.131821 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-9pwb2"] Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.191199 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-inventory\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.191298 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.191499 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgmzh\" (UniqueName: \"kubernetes.io/projected/37b4d7f7-90d4-466c-92a1-dcd83eba1306-kube-api-access-kgmzh\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.191536 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.191563 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ceph\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.293707 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-inventory\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.293899 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.294167 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgmzh\" (UniqueName: \"kubernetes.io/projected/37b4d7f7-90d4-466c-92a1-dcd83eba1306-kube-api-access-kgmzh\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.294232 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.294278 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ceph\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.300630 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-inventory\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.300700 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ceph\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.300734 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.302083 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.319710 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgmzh\" (UniqueName: \"kubernetes.io/projected/37b4d7f7-90d4-466c-92a1-dcd83eba1306-kube-api-access-kgmzh\") pod \"bootstrap-openstack-openstack-cell1-9pwb2\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.448187 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.953967 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-9pwb2"] Nov 24 10:45:47 crc kubenswrapper[4944]: I1124 10:45:47.975868 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" event={"ID":"37b4d7f7-90d4-466c-92a1-dcd83eba1306","Type":"ContainerStarted","Data":"26697f25df64fc820a4d6dbb9c0b16632818e5715877a8760806e78a72b56ed6"} Nov 24 10:45:49 crc kubenswrapper[4944]: I1124 10:45:49.002035 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" event={"ID":"37b4d7f7-90d4-466c-92a1-dcd83eba1306","Type":"ContainerStarted","Data":"42d6350e8e56c65eb4e5a18a5876d6527c3443d4833160b47bc8c2b6d73257a3"} Nov 24 10:45:49 crc kubenswrapper[4944]: I1124 10:45:49.024124 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" podStartSLOduration=1.544194599 podStartE2EDuration="2.024102208s" podCreationTimestamp="2025-11-24 10:45:47 +0000 UTC" firstStartedPulling="2025-11-24 10:45:47.96016351 +0000 UTC m=+6808.494603972" lastFinishedPulling="2025-11-24 10:45:48.440071119 +0000 UTC m=+6808.974511581" observedRunningTime="2025-11-24 10:45:49.016368692 +0000 UTC m=+6809.550809144" watchObservedRunningTime="2025-11-24 10:45:49.024102208 +0000 UTC m=+6809.558542680" Nov 24 10:45:53 crc kubenswrapper[4944]: I1124 10:45:53.549066 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:45:53 crc kubenswrapper[4944]: I1124 10:45:53.549557 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:46:23 crc kubenswrapper[4944]: I1124 10:46:23.549023 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:46:23 crc kubenswrapper[4944]: I1124 10:46:23.549620 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.302365 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7h4j4"] Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.305960 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.315167 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7h4j4"] Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.403092 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-utilities\") pod \"redhat-marketplace-7h4j4\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.403163 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75tgz\" (UniqueName: \"kubernetes.io/projected/e5258e86-277c-421e-8be1-448bbcb2978a-kube-api-access-75tgz\") pod \"redhat-marketplace-7h4j4\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.403193 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-catalog-content\") pod \"redhat-marketplace-7h4j4\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.505213 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-utilities\") pod \"redhat-marketplace-7h4j4\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.505548 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75tgz\" (UniqueName: \"kubernetes.io/projected/e5258e86-277c-421e-8be1-448bbcb2978a-kube-api-access-75tgz\") pod \"redhat-marketplace-7h4j4\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.505568 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-catalog-content\") pod \"redhat-marketplace-7h4j4\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.505931 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-utilities\") pod \"redhat-marketplace-7h4j4\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.505996 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-catalog-content\") pod \"redhat-marketplace-7h4j4\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.533700 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75tgz\" (UniqueName: \"kubernetes.io/projected/e5258e86-277c-421e-8be1-448bbcb2978a-kube-api-access-75tgz\") pod \"redhat-marketplace-7h4j4\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:44 crc kubenswrapper[4944]: I1124 10:46:44.659534 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:45 crc kubenswrapper[4944]: I1124 10:46:45.138666 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7h4j4"] Nov 24 10:46:45 crc kubenswrapper[4944]: I1124 10:46:45.524770 4944 generic.go:334] "Generic (PLEG): container finished" podID="e5258e86-277c-421e-8be1-448bbcb2978a" containerID="c032715a99ad215db5bf7cc1537f8608791e84cc4b73325fbec80b2eaf2ddba0" exitCode=0 Nov 24 10:46:45 crc kubenswrapper[4944]: I1124 10:46:45.524914 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h4j4" event={"ID":"e5258e86-277c-421e-8be1-448bbcb2978a","Type":"ContainerDied","Data":"c032715a99ad215db5bf7cc1537f8608791e84cc4b73325fbec80b2eaf2ddba0"} Nov 24 10:46:45 crc kubenswrapper[4944]: I1124 10:46:45.525109 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h4j4" event={"ID":"e5258e86-277c-421e-8be1-448bbcb2978a","Type":"ContainerStarted","Data":"5eaaee08145671acf50cde4eb31139bcd9e527f7fb70f49d56d35529baecb980"} Nov 24 10:46:47 crc kubenswrapper[4944]: I1124 10:46:47.547478 4944 generic.go:334] "Generic (PLEG): container finished" podID="e5258e86-277c-421e-8be1-448bbcb2978a" containerID="523a4b01ee33a1e1b2a01aa52bcb61ff07d4a8fc45bc710a2bc58068e0133100" exitCode=0 Nov 24 10:46:47 crc kubenswrapper[4944]: I1124 10:46:47.547517 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h4j4" event={"ID":"e5258e86-277c-421e-8be1-448bbcb2978a","Type":"ContainerDied","Data":"523a4b01ee33a1e1b2a01aa52bcb61ff07d4a8fc45bc710a2bc58068e0133100"} Nov 24 10:46:48 crc kubenswrapper[4944]: I1124 10:46:48.558222 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h4j4" event={"ID":"e5258e86-277c-421e-8be1-448bbcb2978a","Type":"ContainerStarted","Data":"1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e"} Nov 24 10:46:48 crc kubenswrapper[4944]: I1124 10:46:48.585947 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7h4j4" podStartSLOduration=2.063267133 podStartE2EDuration="4.585928732s" podCreationTimestamp="2025-11-24 10:46:44 +0000 UTC" firstStartedPulling="2025-11-24 10:46:45.526297189 +0000 UTC m=+6866.060737651" lastFinishedPulling="2025-11-24 10:46:48.048958788 +0000 UTC m=+6868.583399250" observedRunningTime="2025-11-24 10:46:48.574100536 +0000 UTC m=+6869.108541038" watchObservedRunningTime="2025-11-24 10:46:48.585928732 +0000 UTC m=+6869.120369194" Nov 24 10:46:53 crc kubenswrapper[4944]: I1124 10:46:53.548757 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:46:53 crc kubenswrapper[4944]: I1124 10:46:53.549177 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:46:53 crc kubenswrapper[4944]: I1124 10:46:53.549227 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 10:46:53 crc kubenswrapper[4944]: I1124 10:46:53.550074 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7a707b8a170f6fd3b0b500f93173c046a7d3db862dd0277d05c759ca4d9c9eb4"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 10:46:53 crc kubenswrapper[4944]: I1124 10:46:53.550126 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://7a707b8a170f6fd3b0b500f93173c046a7d3db862dd0277d05c759ca4d9c9eb4" gracePeriod=600 Nov 24 10:46:54 crc kubenswrapper[4944]: I1124 10:46:54.614369 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="7a707b8a170f6fd3b0b500f93173c046a7d3db862dd0277d05c759ca4d9c9eb4" exitCode=0 Nov 24 10:46:54 crc kubenswrapper[4944]: I1124 10:46:54.614454 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"7a707b8a170f6fd3b0b500f93173c046a7d3db862dd0277d05c759ca4d9c9eb4"} Nov 24 10:46:54 crc kubenswrapper[4944]: I1124 10:46:54.614986 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053"} Nov 24 10:46:54 crc kubenswrapper[4944]: I1124 10:46:54.615019 4944 scope.go:117] "RemoveContainer" containerID="ea700c5badec10e5f68655544b9f345e7d36e45b4d9ea643f3817dd5f22d09f4" Nov 24 10:46:54 crc kubenswrapper[4944]: I1124 10:46:54.660427 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:54 crc kubenswrapper[4944]: I1124 10:46:54.660484 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:54 crc kubenswrapper[4944]: I1124 10:46:54.715762 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:55 crc kubenswrapper[4944]: I1124 10:46:55.684778 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:55 crc kubenswrapper[4944]: I1124 10:46:55.737232 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7h4j4"] Nov 24 10:46:57 crc kubenswrapper[4944]: I1124 10:46:57.647626 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7h4j4" podUID="e5258e86-277c-421e-8be1-448bbcb2978a" containerName="registry-server" containerID="cri-o://1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e" gracePeriod=2 Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.140210 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.209402 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-catalog-content\") pod \"e5258e86-277c-421e-8be1-448bbcb2978a\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.209559 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75tgz\" (UniqueName: \"kubernetes.io/projected/e5258e86-277c-421e-8be1-448bbcb2978a-kube-api-access-75tgz\") pod \"e5258e86-277c-421e-8be1-448bbcb2978a\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.209587 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-utilities\") pod \"e5258e86-277c-421e-8be1-448bbcb2978a\" (UID: \"e5258e86-277c-421e-8be1-448bbcb2978a\") " Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.210552 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-utilities" (OuterVolumeSpecName: "utilities") pod "e5258e86-277c-421e-8be1-448bbcb2978a" (UID: "e5258e86-277c-421e-8be1-448bbcb2978a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.211368 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.215268 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5258e86-277c-421e-8be1-448bbcb2978a-kube-api-access-75tgz" (OuterVolumeSpecName: "kube-api-access-75tgz") pod "e5258e86-277c-421e-8be1-448bbcb2978a" (UID: "e5258e86-277c-421e-8be1-448bbcb2978a"). InnerVolumeSpecName "kube-api-access-75tgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.225816 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5258e86-277c-421e-8be1-448bbcb2978a" (UID: "e5258e86-277c-421e-8be1-448bbcb2978a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.314129 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5258e86-277c-421e-8be1-448bbcb2978a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.314174 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75tgz\" (UniqueName: \"kubernetes.io/projected/e5258e86-277c-421e-8be1-448bbcb2978a-kube-api-access-75tgz\") on node \"crc\" DevicePath \"\"" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.658662 4944 generic.go:334] "Generic (PLEG): container finished" podID="e5258e86-277c-421e-8be1-448bbcb2978a" containerID="1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e" exitCode=0 Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.658743 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7h4j4" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.658770 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h4j4" event={"ID":"e5258e86-277c-421e-8be1-448bbcb2978a","Type":"ContainerDied","Data":"1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e"} Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.659299 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7h4j4" event={"ID":"e5258e86-277c-421e-8be1-448bbcb2978a","Type":"ContainerDied","Data":"5eaaee08145671acf50cde4eb31139bcd9e527f7fb70f49d56d35529baecb980"} Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.659347 4944 scope.go:117] "RemoveContainer" containerID="1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.679506 4944 scope.go:117] "RemoveContainer" containerID="523a4b01ee33a1e1b2a01aa52bcb61ff07d4a8fc45bc710a2bc58068e0133100" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.687782 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7h4j4"] Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.696231 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7h4j4"] Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.718563 4944 scope.go:117] "RemoveContainer" containerID="c032715a99ad215db5bf7cc1537f8608791e84cc4b73325fbec80b2eaf2ddba0" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.762346 4944 scope.go:117] "RemoveContainer" containerID="1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e" Nov 24 10:46:58 crc kubenswrapper[4944]: E1124 10:46:58.763207 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e\": container with ID starting with 1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e not found: ID does not exist" containerID="1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.763245 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e"} err="failed to get container status \"1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e\": rpc error: code = NotFound desc = could not find container \"1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e\": container with ID starting with 1dcd29c8963db1b0c3ab888fddb435ac0fcf6c90e4ca022cc7d294b1c7b7b86e not found: ID does not exist" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.763271 4944 scope.go:117] "RemoveContainer" containerID="523a4b01ee33a1e1b2a01aa52bcb61ff07d4a8fc45bc710a2bc58068e0133100" Nov 24 10:46:58 crc kubenswrapper[4944]: E1124 10:46:58.763687 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"523a4b01ee33a1e1b2a01aa52bcb61ff07d4a8fc45bc710a2bc58068e0133100\": container with ID starting with 523a4b01ee33a1e1b2a01aa52bcb61ff07d4a8fc45bc710a2bc58068e0133100 not found: ID does not exist" containerID="523a4b01ee33a1e1b2a01aa52bcb61ff07d4a8fc45bc710a2bc58068e0133100" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.763714 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"523a4b01ee33a1e1b2a01aa52bcb61ff07d4a8fc45bc710a2bc58068e0133100"} err="failed to get container status \"523a4b01ee33a1e1b2a01aa52bcb61ff07d4a8fc45bc710a2bc58068e0133100\": rpc error: code = NotFound desc = could not find container \"523a4b01ee33a1e1b2a01aa52bcb61ff07d4a8fc45bc710a2bc58068e0133100\": container with ID starting with 523a4b01ee33a1e1b2a01aa52bcb61ff07d4a8fc45bc710a2bc58068e0133100 not found: ID does not exist" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.763731 4944 scope.go:117] "RemoveContainer" containerID="c032715a99ad215db5bf7cc1537f8608791e84cc4b73325fbec80b2eaf2ddba0" Nov 24 10:46:58 crc kubenswrapper[4944]: E1124 10:46:58.764059 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c032715a99ad215db5bf7cc1537f8608791e84cc4b73325fbec80b2eaf2ddba0\": container with ID starting with c032715a99ad215db5bf7cc1537f8608791e84cc4b73325fbec80b2eaf2ddba0 not found: ID does not exist" containerID="c032715a99ad215db5bf7cc1537f8608791e84cc4b73325fbec80b2eaf2ddba0" Nov 24 10:46:58 crc kubenswrapper[4944]: I1124 10:46:58.764085 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c032715a99ad215db5bf7cc1537f8608791e84cc4b73325fbec80b2eaf2ddba0"} err="failed to get container status \"c032715a99ad215db5bf7cc1537f8608791e84cc4b73325fbec80b2eaf2ddba0\": rpc error: code = NotFound desc = could not find container \"c032715a99ad215db5bf7cc1537f8608791e84cc4b73325fbec80b2eaf2ddba0\": container with ID starting with c032715a99ad215db5bf7cc1537f8608791e84cc4b73325fbec80b2eaf2ddba0 not found: ID does not exist" Nov 24 10:47:00 crc kubenswrapper[4944]: I1124 10:47:00.288691 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5258e86-277c-421e-8be1-448bbcb2978a" path="/var/lib/kubelet/pods/e5258e86-277c-421e-8be1-448bbcb2978a/volumes" Nov 24 10:48:53 crc kubenswrapper[4944]: I1124 10:48:53.548808 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:48:53 crc kubenswrapper[4944]: I1124 10:48:53.549558 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:49:14 crc kubenswrapper[4944]: I1124 10:49:14.292585 4944 generic.go:334] "Generic (PLEG): container finished" podID="37b4d7f7-90d4-466c-92a1-dcd83eba1306" containerID="42d6350e8e56c65eb4e5a18a5876d6527c3443d4833160b47bc8c2b6d73257a3" exitCode=0 Nov 24 10:49:14 crc kubenswrapper[4944]: I1124 10:49:14.292688 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" event={"ID":"37b4d7f7-90d4-466c-92a1-dcd83eba1306","Type":"ContainerDied","Data":"42d6350e8e56c65eb4e5a18a5876d6527c3443d4833160b47bc8c2b6d73257a3"} Nov 24 10:49:15 crc kubenswrapper[4944]: I1124 10:49:15.818704 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:49:15 crc kubenswrapper[4944]: I1124 10:49:15.960204 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ssh-key\") pod \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " Nov 24 10:49:15 crc kubenswrapper[4944]: I1124 10:49:15.960299 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgmzh\" (UniqueName: \"kubernetes.io/projected/37b4d7f7-90d4-466c-92a1-dcd83eba1306-kube-api-access-kgmzh\") pod \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " Nov 24 10:49:15 crc kubenswrapper[4944]: I1124 10:49:15.960399 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-bootstrap-combined-ca-bundle\") pod \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " Nov 24 10:49:15 crc kubenswrapper[4944]: I1124 10:49:15.960493 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ceph\") pod \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " Nov 24 10:49:15 crc kubenswrapper[4944]: I1124 10:49:15.960517 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-inventory\") pod \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\" (UID: \"37b4d7f7-90d4-466c-92a1-dcd83eba1306\") " Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.012788 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37b4d7f7-90d4-466c-92a1-dcd83eba1306-kube-api-access-kgmzh" (OuterVolumeSpecName: "kube-api-access-kgmzh") pod "37b4d7f7-90d4-466c-92a1-dcd83eba1306" (UID: "37b4d7f7-90d4-466c-92a1-dcd83eba1306"). InnerVolumeSpecName "kube-api-access-kgmzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.013211 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "37b4d7f7-90d4-466c-92a1-dcd83eba1306" (UID: "37b4d7f7-90d4-466c-92a1-dcd83eba1306"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.026433 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ceph" (OuterVolumeSpecName: "ceph") pod "37b4d7f7-90d4-466c-92a1-dcd83eba1306" (UID: "37b4d7f7-90d4-466c-92a1-dcd83eba1306"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.042162 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "37b4d7f7-90d4-466c-92a1-dcd83eba1306" (UID: "37b4d7f7-90d4-466c-92a1-dcd83eba1306"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.066097 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.066137 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.066173 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgmzh\" (UniqueName: \"kubernetes.io/projected/37b4d7f7-90d4-466c-92a1-dcd83eba1306-kube-api-access-kgmzh\") on node \"crc\" DevicePath \"\"" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.066185 4944 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.096265 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-inventory" (OuterVolumeSpecName: "inventory") pod "37b4d7f7-90d4-466c-92a1-dcd83eba1306" (UID: "37b4d7f7-90d4-466c-92a1-dcd83eba1306"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.168729 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37b4d7f7-90d4-466c-92a1-dcd83eba1306-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.315617 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" event={"ID":"37b4d7f7-90d4-466c-92a1-dcd83eba1306","Type":"ContainerDied","Data":"26697f25df64fc820a4d6dbb9c0b16632818e5715877a8760806e78a72b56ed6"} Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.315659 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26697f25df64fc820a4d6dbb9c0b16632818e5715877a8760806e78a72b56ed6" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.315712 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-9pwb2" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.388438 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-7tdq4"] Nov 24 10:49:16 crc kubenswrapper[4944]: E1124 10:49:16.389154 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5258e86-277c-421e-8be1-448bbcb2978a" containerName="registry-server" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.389176 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5258e86-277c-421e-8be1-448bbcb2978a" containerName="registry-server" Nov 24 10:49:16 crc kubenswrapper[4944]: E1124 10:49:16.389199 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b4d7f7-90d4-466c-92a1-dcd83eba1306" containerName="bootstrap-openstack-openstack-cell1" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.389207 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b4d7f7-90d4-466c-92a1-dcd83eba1306" containerName="bootstrap-openstack-openstack-cell1" Nov 24 10:49:16 crc kubenswrapper[4944]: E1124 10:49:16.389217 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5258e86-277c-421e-8be1-448bbcb2978a" containerName="extract-utilities" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.389223 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5258e86-277c-421e-8be1-448bbcb2978a" containerName="extract-utilities" Nov 24 10:49:16 crc kubenswrapper[4944]: E1124 10:49:16.389266 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5258e86-277c-421e-8be1-448bbcb2978a" containerName="extract-content" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.389276 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5258e86-277c-421e-8be1-448bbcb2978a" containerName="extract-content" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.389539 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b4d7f7-90d4-466c-92a1-dcd83eba1306" containerName="bootstrap-openstack-openstack-cell1" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.389558 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5258e86-277c-421e-8be1-448bbcb2978a" containerName="registry-server" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.390565 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.395526 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.395551 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.395826 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.396104 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.408699 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-7tdq4"] Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.578998 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ceph\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.579115 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grbwz\" (UniqueName: \"kubernetes.io/projected/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-kube-api-access-grbwz\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.579139 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-inventory\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.579217 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ssh-key\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.681187 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ceph\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.681304 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grbwz\" (UniqueName: \"kubernetes.io/projected/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-kube-api-access-grbwz\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.681332 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-inventory\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.681423 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ssh-key\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.688628 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ssh-key\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.689596 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ceph\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.690962 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-inventory\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.701439 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grbwz\" (UniqueName: \"kubernetes.io/projected/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-kube-api-access-grbwz\") pod \"download-cache-openstack-openstack-cell1-7tdq4\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:16 crc kubenswrapper[4944]: I1124 10:49:16.711145 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:49:17 crc kubenswrapper[4944]: I1124 10:49:17.272157 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-7tdq4"] Nov 24 10:49:17 crc kubenswrapper[4944]: I1124 10:49:17.275228 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 10:49:17 crc kubenswrapper[4944]: I1124 10:49:17.328708 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" event={"ID":"1b22e4e6-4037-4c1d-a2b1-b428efd1f470","Type":"ContainerStarted","Data":"6b3b22bd68e6d008bad0197c23d3892c3542ca6b135bb6b55d7753a1c2f05f9f"} Nov 24 10:49:18 crc kubenswrapper[4944]: I1124 10:49:18.354757 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" event={"ID":"1b22e4e6-4037-4c1d-a2b1-b428efd1f470","Type":"ContainerStarted","Data":"8a5739f49766781a15f24befd2855a0c328aa4877080282eebe34e9d109995cd"} Nov 24 10:49:18 crc kubenswrapper[4944]: I1124 10:49:18.375702 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" podStartSLOduration=1.972438719 podStartE2EDuration="2.375678172s" podCreationTimestamp="2025-11-24 10:49:16 +0000 UTC" firstStartedPulling="2025-11-24 10:49:17.275019917 +0000 UTC m=+7017.809460379" lastFinishedPulling="2025-11-24 10:49:17.67825937 +0000 UTC m=+7018.212699832" observedRunningTime="2025-11-24 10:49:18.371368005 +0000 UTC m=+7018.905808467" watchObservedRunningTime="2025-11-24 10:49:18.375678172 +0000 UTC m=+7018.910118634" Nov 24 10:49:23 crc kubenswrapper[4944]: I1124 10:49:23.548650 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:49:23 crc kubenswrapper[4944]: I1124 10:49:23.549209 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:49:53 crc kubenswrapper[4944]: I1124 10:49:53.548207 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:49:53 crc kubenswrapper[4944]: I1124 10:49:53.548859 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:49:53 crc kubenswrapper[4944]: I1124 10:49:53.548911 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 10:49:53 crc kubenswrapper[4944]: I1124 10:49:53.549768 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 10:49:53 crc kubenswrapper[4944]: I1124 10:49:53.549841 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" gracePeriod=600 Nov 24 10:49:53 crc kubenswrapper[4944]: E1124 10:49:53.684299 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:49:53 crc kubenswrapper[4944]: I1124 10:49:53.684457 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" exitCode=0 Nov 24 10:49:53 crc kubenswrapper[4944]: I1124 10:49:53.684482 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053"} Nov 24 10:49:53 crc kubenswrapper[4944]: I1124 10:49:53.684510 4944 scope.go:117] "RemoveContainer" containerID="7a707b8a170f6fd3b0b500f93173c046a7d3db862dd0277d05c759ca4d9c9eb4" Nov 24 10:49:54 crc kubenswrapper[4944]: I1124 10:49:54.700517 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:49:54 crc kubenswrapper[4944]: E1124 10:49:54.701390 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:50:07 crc kubenswrapper[4944]: I1124 10:50:07.277284 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:50:07 crc kubenswrapper[4944]: E1124 10:50:07.277992 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:50:20 crc kubenswrapper[4944]: I1124 10:50:20.285483 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:50:20 crc kubenswrapper[4944]: E1124 10:50:20.286812 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:50:34 crc kubenswrapper[4944]: I1124 10:50:34.279192 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:50:34 crc kubenswrapper[4944]: E1124 10:50:34.279969 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:50:47 crc kubenswrapper[4944]: I1124 10:50:47.277475 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:50:47 crc kubenswrapper[4944]: E1124 10:50:47.278087 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:50:52 crc kubenswrapper[4944]: I1124 10:50:52.224621 4944 generic.go:334] "Generic (PLEG): container finished" podID="1b22e4e6-4037-4c1d-a2b1-b428efd1f470" containerID="8a5739f49766781a15f24befd2855a0c328aa4877080282eebe34e9d109995cd" exitCode=0 Nov 24 10:50:52 crc kubenswrapper[4944]: I1124 10:50:52.224727 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" event={"ID":"1b22e4e6-4037-4c1d-a2b1-b428efd1f470","Type":"ContainerDied","Data":"8a5739f49766781a15f24befd2855a0c328aa4877080282eebe34e9d109995cd"} Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.685418 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.835148 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ssh-key\") pod \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.835189 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ceph\") pod \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.835275 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-inventory\") pod \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.835507 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grbwz\" (UniqueName: \"kubernetes.io/projected/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-kube-api-access-grbwz\") pod \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\" (UID: \"1b22e4e6-4037-4c1d-a2b1-b428efd1f470\") " Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.840695 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-kube-api-access-grbwz" (OuterVolumeSpecName: "kube-api-access-grbwz") pod "1b22e4e6-4037-4c1d-a2b1-b428efd1f470" (UID: "1b22e4e6-4037-4c1d-a2b1-b428efd1f470"). InnerVolumeSpecName "kube-api-access-grbwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.841703 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ceph" (OuterVolumeSpecName: "ceph") pod "1b22e4e6-4037-4c1d-a2b1-b428efd1f470" (UID: "1b22e4e6-4037-4c1d-a2b1-b428efd1f470"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.864035 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-inventory" (OuterVolumeSpecName: "inventory") pod "1b22e4e6-4037-4c1d-a2b1-b428efd1f470" (UID: "1b22e4e6-4037-4c1d-a2b1-b428efd1f470"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.876154 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1b22e4e6-4037-4c1d-a2b1-b428efd1f470" (UID: "1b22e4e6-4037-4c1d-a2b1-b428efd1f470"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.937775 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grbwz\" (UniqueName: \"kubernetes.io/projected/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-kube-api-access-grbwz\") on node \"crc\" DevicePath \"\"" Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.937811 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.937822 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:50:53 crc kubenswrapper[4944]: I1124 10:50:53.937830 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b22e4e6-4037-4c1d-a2b1-b428efd1f470-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.250982 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" event={"ID":"1b22e4e6-4037-4c1d-a2b1-b428efd1f470","Type":"ContainerDied","Data":"6b3b22bd68e6d008bad0197c23d3892c3542ca6b135bb6b55d7753a1c2f05f9f"} Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.251043 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b3b22bd68e6d008bad0197c23d3892c3542ca6b135bb6b55d7753a1c2f05f9f" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.251142 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-7tdq4" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.374775 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-sxf2s"] Nov 24 10:50:54 crc kubenswrapper[4944]: E1124 10:50:54.375206 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b22e4e6-4037-4c1d-a2b1-b428efd1f470" containerName="download-cache-openstack-openstack-cell1" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.375226 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b22e4e6-4037-4c1d-a2b1-b428efd1f470" containerName="download-cache-openstack-openstack-cell1" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.375485 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b22e4e6-4037-4c1d-a2b1-b428efd1f470" containerName="download-cache-openstack-openstack-cell1" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.376237 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.378617 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.379536 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.379667 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.379832 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.389185 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-sxf2s"] Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.576366 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gwmd\" (UniqueName: \"kubernetes.io/projected/f09923cd-e53d-4abd-a2e6-1bec30789bbd-kube-api-access-7gwmd\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.576447 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ssh-key\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.576773 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-inventory\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.576962 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ceph\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.678396 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ssh-key\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.678785 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-inventory\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.678846 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ceph\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.678935 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gwmd\" (UniqueName: \"kubernetes.io/projected/f09923cd-e53d-4abd-a2e6-1bec30789bbd-kube-api-access-7gwmd\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.682266 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-inventory\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.694349 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ssh-key\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.694505 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ceph\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.694555 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gwmd\" (UniqueName: \"kubernetes.io/projected/f09923cd-e53d-4abd-a2e6-1bec30789bbd-kube-api-access-7gwmd\") pod \"configure-network-openstack-openstack-cell1-sxf2s\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:54 crc kubenswrapper[4944]: I1124 10:50:54.709662 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:50:55 crc kubenswrapper[4944]: I1124 10:50:55.289384 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-sxf2s"] Nov 24 10:50:55 crc kubenswrapper[4944]: W1124 10:50:55.297047 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf09923cd_e53d_4abd_a2e6_1bec30789bbd.slice/crio-04cdd00118d564eb06aaac08a274be8dab83a22ed349118f1e33dcb811a5e1b8 WatchSource:0}: Error finding container 04cdd00118d564eb06aaac08a274be8dab83a22ed349118f1e33dcb811a5e1b8: Status 404 returned error can't find the container with id 04cdd00118d564eb06aaac08a274be8dab83a22ed349118f1e33dcb811a5e1b8 Nov 24 10:50:56 crc kubenswrapper[4944]: I1124 10:50:56.287602 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" event={"ID":"f09923cd-e53d-4abd-a2e6-1bec30789bbd","Type":"ContainerStarted","Data":"d8c7f2655d520e511ace586132b4c243609e9c756350958486e3aa2fce6bfb70"} Nov 24 10:50:56 crc kubenswrapper[4944]: I1124 10:50:56.288152 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" event={"ID":"f09923cd-e53d-4abd-a2e6-1bec30789bbd","Type":"ContainerStarted","Data":"04cdd00118d564eb06aaac08a274be8dab83a22ed349118f1e33dcb811a5e1b8"} Nov 24 10:50:56 crc kubenswrapper[4944]: I1124 10:50:56.299571 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" podStartSLOduration=1.927706138 podStartE2EDuration="2.299552493s" podCreationTimestamp="2025-11-24 10:50:54 +0000 UTC" firstStartedPulling="2025-11-24 10:50:55.299347431 +0000 UTC m=+7115.833787893" lastFinishedPulling="2025-11-24 10:50:55.671193776 +0000 UTC m=+7116.205634248" observedRunningTime="2025-11-24 10:50:56.294834463 +0000 UTC m=+7116.829274925" watchObservedRunningTime="2025-11-24 10:50:56.299552493 +0000 UTC m=+7116.833992965" Nov 24 10:50:59 crc kubenswrapper[4944]: I1124 10:50:59.276803 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:50:59 crc kubenswrapper[4944]: E1124 10:50:59.277463 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:51:14 crc kubenswrapper[4944]: I1124 10:51:14.277071 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:51:14 crc kubenswrapper[4944]: E1124 10:51:14.278842 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:51:25 crc kubenswrapper[4944]: I1124 10:51:25.277138 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:51:25 crc kubenswrapper[4944]: E1124 10:51:25.277904 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:51:36 crc kubenswrapper[4944]: I1124 10:51:36.278291 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:51:36 crc kubenswrapper[4944]: E1124 10:51:36.279747 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:51:49 crc kubenswrapper[4944]: I1124 10:51:49.277848 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:51:49 crc kubenswrapper[4944]: E1124 10:51:49.278809 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:52:02 crc kubenswrapper[4944]: I1124 10:52:02.277920 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:52:02 crc kubenswrapper[4944]: E1124 10:52:02.279240 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:52:13 crc kubenswrapper[4944]: I1124 10:52:13.277985 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:52:13 crc kubenswrapper[4944]: E1124 10:52:13.279070 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:52:17 crc kubenswrapper[4944]: I1124 10:52:17.110875 4944 generic.go:334] "Generic (PLEG): container finished" podID="f09923cd-e53d-4abd-a2e6-1bec30789bbd" containerID="d8c7f2655d520e511ace586132b4c243609e9c756350958486e3aa2fce6bfb70" exitCode=0 Nov 24 10:52:17 crc kubenswrapper[4944]: I1124 10:52:17.110962 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" event={"ID":"f09923cd-e53d-4abd-a2e6-1bec30789bbd","Type":"ContainerDied","Data":"d8c7f2655d520e511ace586132b4c243609e9c756350958486e3aa2fce6bfb70"} Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.590206 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.761904 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ssh-key\") pod \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.761942 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-inventory\") pod \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.762069 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gwmd\" (UniqueName: \"kubernetes.io/projected/f09923cd-e53d-4abd-a2e6-1bec30789bbd-kube-api-access-7gwmd\") pod \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.762188 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ceph\") pod \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\" (UID: \"f09923cd-e53d-4abd-a2e6-1bec30789bbd\") " Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.767294 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f09923cd-e53d-4abd-a2e6-1bec30789bbd-kube-api-access-7gwmd" (OuterVolumeSpecName: "kube-api-access-7gwmd") pod "f09923cd-e53d-4abd-a2e6-1bec30789bbd" (UID: "f09923cd-e53d-4abd-a2e6-1bec30789bbd"). InnerVolumeSpecName "kube-api-access-7gwmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.768968 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ceph" (OuterVolumeSpecName: "ceph") pod "f09923cd-e53d-4abd-a2e6-1bec30789bbd" (UID: "f09923cd-e53d-4abd-a2e6-1bec30789bbd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.790209 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-inventory" (OuterVolumeSpecName: "inventory") pod "f09923cd-e53d-4abd-a2e6-1bec30789bbd" (UID: "f09923cd-e53d-4abd-a2e6-1bec30789bbd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.790619 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f09923cd-e53d-4abd-a2e6-1bec30789bbd" (UID: "f09923cd-e53d-4abd-a2e6-1bec30789bbd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.864731 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.864918 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.864980 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gwmd\" (UniqueName: \"kubernetes.io/projected/f09923cd-e53d-4abd-a2e6-1bec30789bbd-kube-api-access-7gwmd\") on node \"crc\" DevicePath \"\"" Nov 24 10:52:18 crc kubenswrapper[4944]: I1124 10:52:18.865103 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f09923cd-e53d-4abd-a2e6-1bec30789bbd-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.133791 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" event={"ID":"f09923cd-e53d-4abd-a2e6-1bec30789bbd","Type":"ContainerDied","Data":"04cdd00118d564eb06aaac08a274be8dab83a22ed349118f1e33dcb811a5e1b8"} Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.134161 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04cdd00118d564eb06aaac08a274be8dab83a22ed349118f1e33dcb811a5e1b8" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.133869 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-sxf2s" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.229467 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-bm2rd"] Nov 24 10:52:19 crc kubenswrapper[4944]: E1124 10:52:19.229927 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f09923cd-e53d-4abd-a2e6-1bec30789bbd" containerName="configure-network-openstack-openstack-cell1" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.229945 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f09923cd-e53d-4abd-a2e6-1bec30789bbd" containerName="configure-network-openstack-openstack-cell1" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.230211 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f09923cd-e53d-4abd-a2e6-1bec30789bbd" containerName="configure-network-openstack-openstack-cell1" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.230950 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.233142 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.233158 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.233347 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.233366 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.244713 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-bm2rd"] Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.374860 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-inventory\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.374942 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ssh-key\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.375112 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ceph\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.375373 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx7hv\" (UniqueName: \"kubernetes.io/projected/bd996416-36cc-46b8-8982-78c11da1243c-kube-api-access-wx7hv\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.477113 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ceph\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.477222 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx7hv\" (UniqueName: \"kubernetes.io/projected/bd996416-36cc-46b8-8982-78c11da1243c-kube-api-access-wx7hv\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.477331 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-inventory\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.477400 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ssh-key\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.481477 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ssh-key\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.481645 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-inventory\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.484538 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ceph\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.491992 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx7hv\" (UniqueName: \"kubernetes.io/projected/bd996416-36cc-46b8-8982-78c11da1243c-kube-api-access-wx7hv\") pod \"validate-network-openstack-openstack-cell1-bm2rd\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:19 crc kubenswrapper[4944]: I1124 10:52:19.550250 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:20 crc kubenswrapper[4944]: I1124 10:52:20.055795 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-bm2rd"] Nov 24 10:52:20 crc kubenswrapper[4944]: I1124 10:52:20.156172 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" event={"ID":"bd996416-36cc-46b8-8982-78c11da1243c","Type":"ContainerStarted","Data":"6694e1fd8ef4d8e535c112be5a0901958cc6302cdac0b7bb143e7953937327b6"} Nov 24 10:52:20 crc kubenswrapper[4944]: I1124 10:52:20.481474 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:52:21 crc kubenswrapper[4944]: I1124 10:52:21.166763 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" event={"ID":"bd996416-36cc-46b8-8982-78c11da1243c","Type":"ContainerStarted","Data":"3f160b6e35037be1a627b52764d416a61b22ac70965572dce138bad5445773ba"} Nov 24 10:52:21 crc kubenswrapper[4944]: I1124 10:52:21.186852 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" podStartSLOduration=1.775106825 podStartE2EDuration="2.186836238s" podCreationTimestamp="2025-11-24 10:52:19 +0000 UTC" firstStartedPulling="2025-11-24 10:52:20.066901501 +0000 UTC m=+7200.601341963" lastFinishedPulling="2025-11-24 10:52:20.478630914 +0000 UTC m=+7201.013071376" observedRunningTime="2025-11-24 10:52:21.182654036 +0000 UTC m=+7201.717094518" watchObservedRunningTime="2025-11-24 10:52:21.186836238 +0000 UTC m=+7201.721276700" Nov 24 10:52:24 crc kubenswrapper[4944]: I1124 10:52:24.277303 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:52:24 crc kubenswrapper[4944]: E1124 10:52:24.278264 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:52:25 crc kubenswrapper[4944]: I1124 10:52:25.212589 4944 generic.go:334] "Generic (PLEG): container finished" podID="bd996416-36cc-46b8-8982-78c11da1243c" containerID="3f160b6e35037be1a627b52764d416a61b22ac70965572dce138bad5445773ba" exitCode=0 Nov 24 10:52:25 crc kubenswrapper[4944]: I1124 10:52:25.212789 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" event={"ID":"bd996416-36cc-46b8-8982-78c11da1243c","Type":"ContainerDied","Data":"3f160b6e35037be1a627b52764d416a61b22ac70965572dce138bad5445773ba"} Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.629550 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.731028 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx7hv\" (UniqueName: \"kubernetes.io/projected/bd996416-36cc-46b8-8982-78c11da1243c-kube-api-access-wx7hv\") pod \"bd996416-36cc-46b8-8982-78c11da1243c\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.731181 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-inventory\") pod \"bd996416-36cc-46b8-8982-78c11da1243c\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.731240 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ssh-key\") pod \"bd996416-36cc-46b8-8982-78c11da1243c\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.731305 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ceph\") pod \"bd996416-36cc-46b8-8982-78c11da1243c\" (UID: \"bd996416-36cc-46b8-8982-78c11da1243c\") " Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.737906 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd996416-36cc-46b8-8982-78c11da1243c-kube-api-access-wx7hv" (OuterVolumeSpecName: "kube-api-access-wx7hv") pod "bd996416-36cc-46b8-8982-78c11da1243c" (UID: "bd996416-36cc-46b8-8982-78c11da1243c"). InnerVolumeSpecName "kube-api-access-wx7hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.738533 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ceph" (OuterVolumeSpecName: "ceph") pod "bd996416-36cc-46b8-8982-78c11da1243c" (UID: "bd996416-36cc-46b8-8982-78c11da1243c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.767874 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-inventory" (OuterVolumeSpecName: "inventory") pod "bd996416-36cc-46b8-8982-78c11da1243c" (UID: "bd996416-36cc-46b8-8982-78c11da1243c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.773435 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bd996416-36cc-46b8-8982-78c11da1243c" (UID: "bd996416-36cc-46b8-8982-78c11da1243c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.833972 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx7hv\" (UniqueName: \"kubernetes.io/projected/bd996416-36cc-46b8-8982-78c11da1243c-kube-api-access-wx7hv\") on node \"crc\" DevicePath \"\"" Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.834009 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.834020 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:52:26 crc kubenswrapper[4944]: I1124 10:52:26.834029 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bd996416-36cc-46b8-8982-78c11da1243c-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.231167 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" event={"ID":"bd996416-36cc-46b8-8982-78c11da1243c","Type":"ContainerDied","Data":"6694e1fd8ef4d8e535c112be5a0901958cc6302cdac0b7bb143e7953937327b6"} Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.231208 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6694e1fd8ef4d8e535c112be5a0901958cc6302cdac0b7bb143e7953937327b6" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.231208 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-bm2rd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.293706 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-89pgd"] Nov 24 10:52:27 crc kubenswrapper[4944]: E1124 10:52:27.294272 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd996416-36cc-46b8-8982-78c11da1243c" containerName="validate-network-openstack-openstack-cell1" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.294294 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd996416-36cc-46b8-8982-78c11da1243c" containerName="validate-network-openstack-openstack-cell1" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.294596 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd996416-36cc-46b8-8982-78c11da1243c" containerName="validate-network-openstack-openstack-cell1" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.295614 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.298551 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.299033 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.299346 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.299695 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.316263 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-89pgd"] Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.451097 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ceph\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.451211 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-inventory\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.451250 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fkcf\" (UniqueName: \"kubernetes.io/projected/3317708a-6513-47e2-b70f-b4137e7e25d1-kube-api-access-7fkcf\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.451334 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ssh-key\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.510800 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h7nbj"] Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.513592 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.526111 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h7nbj"] Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.553557 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ssh-key\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.553687 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ceph\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.553754 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-inventory\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.553791 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fkcf\" (UniqueName: \"kubernetes.io/projected/3317708a-6513-47e2-b70f-b4137e7e25d1-kube-api-access-7fkcf\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.557431 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ssh-key\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.557607 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ceph\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.558412 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-inventory\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.570504 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fkcf\" (UniqueName: \"kubernetes.io/projected/3317708a-6513-47e2-b70f-b4137e7e25d1-kube-api-access-7fkcf\") pod \"install-os-openstack-openstack-cell1-89pgd\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.613195 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.655681 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npslm\" (UniqueName: \"kubernetes.io/projected/93aeed16-88c6-4027-9efa-94c1286469c5-kube-api-access-npslm\") pod \"redhat-operators-h7nbj\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.655961 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-catalog-content\") pod \"redhat-operators-h7nbj\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.655986 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-utilities\") pod \"redhat-operators-h7nbj\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.758524 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npslm\" (UniqueName: \"kubernetes.io/projected/93aeed16-88c6-4027-9efa-94c1286469c5-kube-api-access-npslm\") pod \"redhat-operators-h7nbj\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.758603 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-catalog-content\") pod \"redhat-operators-h7nbj\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.758623 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-utilities\") pod \"redhat-operators-h7nbj\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.759105 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-catalog-content\") pod \"redhat-operators-h7nbj\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.759152 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-utilities\") pod \"redhat-operators-h7nbj\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.777184 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npslm\" (UniqueName: \"kubernetes.io/projected/93aeed16-88c6-4027-9efa-94c1286469c5-kube-api-access-npslm\") pod \"redhat-operators-h7nbj\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:27 crc kubenswrapper[4944]: I1124 10:52:27.841013 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:28 crc kubenswrapper[4944]: I1124 10:52:28.126932 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-89pgd"] Nov 24 10:52:28 crc kubenswrapper[4944]: I1124 10:52:28.241263 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-89pgd" event={"ID":"3317708a-6513-47e2-b70f-b4137e7e25d1","Type":"ContainerStarted","Data":"0b4fbc4739b629af20f73b3cff18096fc05240dee561e7f1043be841d2d33554"} Nov 24 10:52:28 crc kubenswrapper[4944]: W1124 10:52:28.283946 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93aeed16_88c6_4027_9efa_94c1286469c5.slice/crio-be6063d2854f2a38d9afc52fb2fd1a22745846c36d05ed1036df36a349c7863d WatchSource:0}: Error finding container be6063d2854f2a38d9afc52fb2fd1a22745846c36d05ed1036df36a349c7863d: Status 404 returned error can't find the container with id be6063d2854f2a38d9afc52fb2fd1a22745846c36d05ed1036df36a349c7863d Nov 24 10:52:28 crc kubenswrapper[4944]: I1124 10:52:28.290482 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h7nbj"] Nov 24 10:52:29 crc kubenswrapper[4944]: I1124 10:52:29.252288 4944 generic.go:334] "Generic (PLEG): container finished" podID="93aeed16-88c6-4027-9efa-94c1286469c5" containerID="acaa5967079f42908c966a81e230f5dcbf36ca9400dbdaf01f18d598d42350f6" exitCode=0 Nov 24 10:52:29 crc kubenswrapper[4944]: I1124 10:52:29.252364 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7nbj" event={"ID":"93aeed16-88c6-4027-9efa-94c1286469c5","Type":"ContainerDied","Data":"acaa5967079f42908c966a81e230f5dcbf36ca9400dbdaf01f18d598d42350f6"} Nov 24 10:52:29 crc kubenswrapper[4944]: I1124 10:52:29.252723 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7nbj" event={"ID":"93aeed16-88c6-4027-9efa-94c1286469c5","Type":"ContainerStarted","Data":"be6063d2854f2a38d9afc52fb2fd1a22745846c36d05ed1036df36a349c7863d"} Nov 24 10:52:29 crc kubenswrapper[4944]: I1124 10:52:29.254805 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-89pgd" event={"ID":"3317708a-6513-47e2-b70f-b4137e7e25d1","Type":"ContainerStarted","Data":"bbc4f13846541a09bbfd7a55ad449e368f950b468aaaabbf5548e8ea235615a8"} Nov 24 10:52:29 crc kubenswrapper[4944]: I1124 10:52:29.292452 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-89pgd" podStartSLOduration=1.8411884010000001 podStartE2EDuration="2.29242316s" podCreationTimestamp="2025-11-24 10:52:27 +0000 UTC" firstStartedPulling="2025-11-24 10:52:28.130373695 +0000 UTC m=+7208.664814157" lastFinishedPulling="2025-11-24 10:52:28.581608454 +0000 UTC m=+7209.116048916" observedRunningTime="2025-11-24 10:52:29.290084456 +0000 UTC m=+7209.824524938" watchObservedRunningTime="2025-11-24 10:52:29.29242316 +0000 UTC m=+7209.826863622" Nov 24 10:52:30 crc kubenswrapper[4944]: I1124 10:52:30.266036 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7nbj" event={"ID":"93aeed16-88c6-4027-9efa-94c1286469c5","Type":"ContainerStarted","Data":"efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e"} Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.202053 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wr5wb"] Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.205737 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.212977 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wr5wb"] Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.286372 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-utilities\") pod \"certified-operators-wr5wb\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.286433 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-catalog-content\") pod \"certified-operators-wr5wb\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.286594 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8vc5\" (UniqueName: \"kubernetes.io/projected/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-kube-api-access-w8vc5\") pod \"certified-operators-wr5wb\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.388478 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-catalog-content\") pod \"certified-operators-wr5wb\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.388969 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8vc5\" (UniqueName: \"kubernetes.io/projected/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-kube-api-access-w8vc5\") pod \"certified-operators-wr5wb\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.389105 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-utilities\") pod \"certified-operators-wr5wb\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.389254 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-catalog-content\") pod \"certified-operators-wr5wb\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.389609 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-utilities\") pod \"certified-operators-wr5wb\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.424007 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8vc5\" (UniqueName: \"kubernetes.io/projected/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-kube-api-access-w8vc5\") pod \"certified-operators-wr5wb\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:33 crc kubenswrapper[4944]: I1124 10:52:33.541910 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:34 crc kubenswrapper[4944]: I1124 10:52:34.094333 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wr5wb"] Nov 24 10:52:34 crc kubenswrapper[4944]: W1124 10:52:34.108609 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa6876f3_2ad9_4ac9_ac2a_cac2aef4f667.slice/crio-1c21e7aec65ef0b87527f3e243ddf57b07126ed38502248e5a86b8666e02f3d7 WatchSource:0}: Error finding container 1c21e7aec65ef0b87527f3e243ddf57b07126ed38502248e5a86b8666e02f3d7: Status 404 returned error can't find the container with id 1c21e7aec65ef0b87527f3e243ddf57b07126ed38502248e5a86b8666e02f3d7 Nov 24 10:52:34 crc kubenswrapper[4944]: I1124 10:52:34.368004 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wr5wb" event={"ID":"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667","Type":"ContainerStarted","Data":"1c21e7aec65ef0b87527f3e243ddf57b07126ed38502248e5a86b8666e02f3d7"} Nov 24 10:52:34 crc kubenswrapper[4944]: I1124 10:52:34.397465 4944 generic.go:334] "Generic (PLEG): container finished" podID="93aeed16-88c6-4027-9efa-94c1286469c5" containerID="efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e" exitCode=0 Nov 24 10:52:34 crc kubenswrapper[4944]: I1124 10:52:34.397518 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7nbj" event={"ID":"93aeed16-88c6-4027-9efa-94c1286469c5","Type":"ContainerDied","Data":"efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e"} Nov 24 10:52:35 crc kubenswrapper[4944]: I1124 10:52:35.407910 4944 generic.go:334] "Generic (PLEG): container finished" podID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" containerID="25fe7dc1c502433f52bf5f530ac2e6d1accaca9dca42e034617dcdc803fa6565" exitCode=0 Nov 24 10:52:35 crc kubenswrapper[4944]: I1124 10:52:35.408096 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wr5wb" event={"ID":"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667","Type":"ContainerDied","Data":"25fe7dc1c502433f52bf5f530ac2e6d1accaca9dca42e034617dcdc803fa6565"} Nov 24 10:52:35 crc kubenswrapper[4944]: I1124 10:52:35.411223 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7nbj" event={"ID":"93aeed16-88c6-4027-9efa-94c1286469c5","Type":"ContainerStarted","Data":"c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18"} Nov 24 10:52:35 crc kubenswrapper[4944]: I1124 10:52:35.448280 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h7nbj" podStartSLOduration=2.826743168 podStartE2EDuration="8.448265196s" podCreationTimestamp="2025-11-24 10:52:27 +0000 UTC" firstStartedPulling="2025-11-24 10:52:29.254487345 +0000 UTC m=+7209.788927807" lastFinishedPulling="2025-11-24 10:52:34.876009373 +0000 UTC m=+7215.410449835" observedRunningTime="2025-11-24 10:52:35.446577243 +0000 UTC m=+7215.981017715" watchObservedRunningTime="2025-11-24 10:52:35.448265196 +0000 UTC m=+7215.982705658" Nov 24 10:52:37 crc kubenswrapper[4944]: I1124 10:52:37.434039 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wr5wb" event={"ID":"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667","Type":"ContainerStarted","Data":"ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47"} Nov 24 10:52:37 crc kubenswrapper[4944]: I1124 10:52:37.841394 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:37 crc kubenswrapper[4944]: I1124 10:52:37.841655 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:38 crc kubenswrapper[4944]: I1124 10:52:38.277032 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:52:38 crc kubenswrapper[4944]: E1124 10:52:38.277646 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:52:38 crc kubenswrapper[4944]: I1124 10:52:38.444300 4944 generic.go:334] "Generic (PLEG): container finished" podID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" containerID="ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47" exitCode=0 Nov 24 10:52:38 crc kubenswrapper[4944]: I1124 10:52:38.445116 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wr5wb" event={"ID":"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667","Type":"ContainerDied","Data":"ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47"} Nov 24 10:52:38 crc kubenswrapper[4944]: I1124 10:52:38.898695 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h7nbj" podUID="93aeed16-88c6-4027-9efa-94c1286469c5" containerName="registry-server" probeResult="failure" output=< Nov 24 10:52:38 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 10:52:38 crc kubenswrapper[4944]: > Nov 24 10:52:39 crc kubenswrapper[4944]: I1124 10:52:39.454537 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wr5wb" event={"ID":"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667","Type":"ContainerStarted","Data":"59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3"} Nov 24 10:52:39 crc kubenswrapper[4944]: I1124 10:52:39.475635 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wr5wb" podStartSLOduration=4.047057485 podStartE2EDuration="7.475617219s" podCreationTimestamp="2025-11-24 10:52:32 +0000 UTC" firstStartedPulling="2025-11-24 10:52:35.409809735 +0000 UTC m=+7215.944250197" lastFinishedPulling="2025-11-24 10:52:38.838369469 +0000 UTC m=+7219.372809931" observedRunningTime="2025-11-24 10:52:39.47062662 +0000 UTC m=+7220.005067082" watchObservedRunningTime="2025-11-24 10:52:39.475617219 +0000 UTC m=+7220.010057681" Nov 24 10:52:43 crc kubenswrapper[4944]: I1124 10:52:43.543120 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:43 crc kubenswrapper[4944]: I1124 10:52:43.543769 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:43 crc kubenswrapper[4944]: I1124 10:52:43.589688 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:44 crc kubenswrapper[4944]: I1124 10:52:44.550815 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:44 crc kubenswrapper[4944]: I1124 10:52:44.600164 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wr5wb"] Nov 24 10:52:46 crc kubenswrapper[4944]: I1124 10:52:46.519756 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wr5wb" podUID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" containerName="registry-server" containerID="cri-o://59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3" gracePeriod=2 Nov 24 10:52:46 crc kubenswrapper[4944]: E1124 10:52:46.593605 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa6876f3_2ad9_4ac9_ac2a_cac2aef4f667.slice/crio-59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3.scope\": RecentStats: unable to find data in memory cache]" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.156917 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.205210 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-utilities\") pod \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.205455 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8vc5\" (UniqueName: \"kubernetes.io/projected/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-kube-api-access-w8vc5\") pod \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.205627 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-catalog-content\") pod \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\" (UID: \"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667\") " Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.207015 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-utilities" (OuterVolumeSpecName: "utilities") pod "aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" (UID: "aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.212887 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-kube-api-access-w8vc5" (OuterVolumeSpecName: "kube-api-access-w8vc5") pod "aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" (UID: "aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667"). InnerVolumeSpecName "kube-api-access-w8vc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.253258 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" (UID: "aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.308655 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.308696 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.308712 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8vc5\" (UniqueName: \"kubernetes.io/projected/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667-kube-api-access-w8vc5\") on node \"crc\" DevicePath \"\"" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.546928 4944 generic.go:334] "Generic (PLEG): container finished" podID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" containerID="59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3" exitCode=0 Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.546975 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wr5wb" event={"ID":"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667","Type":"ContainerDied","Data":"59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3"} Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.547006 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wr5wb" event={"ID":"aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667","Type":"ContainerDied","Data":"1c21e7aec65ef0b87527f3e243ddf57b07126ed38502248e5a86b8666e02f3d7"} Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.547028 4944 scope.go:117] "RemoveContainer" containerID="59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.547223 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wr5wb" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.583483 4944 scope.go:117] "RemoveContainer" containerID="ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.587738 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wr5wb"] Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.599713 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wr5wb"] Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.617287 4944 scope.go:117] "RemoveContainer" containerID="25fe7dc1c502433f52bf5f530ac2e6d1accaca9dca42e034617dcdc803fa6565" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.658563 4944 scope.go:117] "RemoveContainer" containerID="59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3" Nov 24 10:52:47 crc kubenswrapper[4944]: E1124 10:52:47.658942 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3\": container with ID starting with 59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3 not found: ID does not exist" containerID="59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.659074 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3"} err="failed to get container status \"59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3\": rpc error: code = NotFound desc = could not find container \"59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3\": container with ID starting with 59a1d061799980ddc5ca639754430809c8ac42d2aadc95e976c0450ba60a3ed3 not found: ID does not exist" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.659198 4944 scope.go:117] "RemoveContainer" containerID="ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47" Nov 24 10:52:47 crc kubenswrapper[4944]: E1124 10:52:47.660474 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47\": container with ID starting with ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47 not found: ID does not exist" containerID="ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.660500 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47"} err="failed to get container status \"ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47\": rpc error: code = NotFound desc = could not find container \"ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47\": container with ID starting with ad4b4b992d22711b972a912552a2763b775c698de300c9feb8c1efa12ba43d47 not found: ID does not exist" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.660543 4944 scope.go:117] "RemoveContainer" containerID="25fe7dc1c502433f52bf5f530ac2e6d1accaca9dca42e034617dcdc803fa6565" Nov 24 10:52:47 crc kubenswrapper[4944]: E1124 10:52:47.660817 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25fe7dc1c502433f52bf5f530ac2e6d1accaca9dca42e034617dcdc803fa6565\": container with ID starting with 25fe7dc1c502433f52bf5f530ac2e6d1accaca9dca42e034617dcdc803fa6565 not found: ID does not exist" containerID="25fe7dc1c502433f52bf5f530ac2e6d1accaca9dca42e034617dcdc803fa6565" Nov 24 10:52:47 crc kubenswrapper[4944]: I1124 10:52:47.660843 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25fe7dc1c502433f52bf5f530ac2e6d1accaca9dca42e034617dcdc803fa6565"} err="failed to get container status \"25fe7dc1c502433f52bf5f530ac2e6d1accaca9dca42e034617dcdc803fa6565\": rpc error: code = NotFound desc = could not find container \"25fe7dc1c502433f52bf5f530ac2e6d1accaca9dca42e034617dcdc803fa6565\": container with ID starting with 25fe7dc1c502433f52bf5f530ac2e6d1accaca9dca42e034617dcdc803fa6565 not found: ID does not exist" Nov 24 10:52:48 crc kubenswrapper[4944]: I1124 10:52:48.289353 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" path="/var/lib/kubelet/pods/aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667/volumes" Nov 24 10:52:48 crc kubenswrapper[4944]: I1124 10:52:48.890547 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h7nbj" podUID="93aeed16-88c6-4027-9efa-94c1286469c5" containerName="registry-server" probeResult="failure" output=< Nov 24 10:52:48 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 10:52:48 crc kubenswrapper[4944]: > Nov 24 10:52:51 crc kubenswrapper[4944]: I1124 10:52:51.277250 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:52:51 crc kubenswrapper[4944]: E1124 10:52:51.278074 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:52:57 crc kubenswrapper[4944]: I1124 10:52:57.888823 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:57 crc kubenswrapper[4944]: I1124 10:52:57.955391 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:52:58 crc kubenswrapper[4944]: I1124 10:52:58.704282 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h7nbj"] Nov 24 10:52:59 crc kubenswrapper[4944]: I1124 10:52:59.661612 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h7nbj" podUID="93aeed16-88c6-4027-9efa-94c1286469c5" containerName="registry-server" containerID="cri-o://c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18" gracePeriod=2 Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.158562 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.270531 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-catalog-content\") pod \"93aeed16-88c6-4027-9efa-94c1286469c5\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.271089 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npslm\" (UniqueName: \"kubernetes.io/projected/93aeed16-88c6-4027-9efa-94c1286469c5-kube-api-access-npslm\") pod \"93aeed16-88c6-4027-9efa-94c1286469c5\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.271244 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-utilities\") pod \"93aeed16-88c6-4027-9efa-94c1286469c5\" (UID: \"93aeed16-88c6-4027-9efa-94c1286469c5\") " Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.272082 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-utilities" (OuterVolumeSpecName: "utilities") pod "93aeed16-88c6-4027-9efa-94c1286469c5" (UID: "93aeed16-88c6-4027-9efa-94c1286469c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.276362 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93aeed16-88c6-4027-9efa-94c1286469c5-kube-api-access-npslm" (OuterVolumeSpecName: "kube-api-access-npslm") pod "93aeed16-88c6-4027-9efa-94c1286469c5" (UID: "93aeed16-88c6-4027-9efa-94c1286469c5"). InnerVolumeSpecName "kube-api-access-npslm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.374862 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npslm\" (UniqueName: \"kubernetes.io/projected/93aeed16-88c6-4027-9efa-94c1286469c5-kube-api-access-npslm\") on node \"crc\" DevicePath \"\"" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.374909 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.376476 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93aeed16-88c6-4027-9efa-94c1286469c5" (UID: "93aeed16-88c6-4027-9efa-94c1286469c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.477039 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93aeed16-88c6-4027-9efa-94c1286469c5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.674655 4944 generic.go:334] "Generic (PLEG): container finished" podID="93aeed16-88c6-4027-9efa-94c1286469c5" containerID="c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18" exitCode=0 Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.674721 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h7nbj" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.674714 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7nbj" event={"ID":"93aeed16-88c6-4027-9efa-94c1286469c5","Type":"ContainerDied","Data":"c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18"} Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.674878 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h7nbj" event={"ID":"93aeed16-88c6-4027-9efa-94c1286469c5","Type":"ContainerDied","Data":"be6063d2854f2a38d9afc52fb2fd1a22745846c36d05ed1036df36a349c7863d"} Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.674907 4944 scope.go:117] "RemoveContainer" containerID="c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.708276 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h7nbj"] Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.712661 4944 scope.go:117] "RemoveContainer" containerID="efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.718342 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h7nbj"] Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.744899 4944 scope.go:117] "RemoveContainer" containerID="acaa5967079f42908c966a81e230f5dcbf36ca9400dbdaf01f18d598d42350f6" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.787667 4944 scope.go:117] "RemoveContainer" containerID="c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18" Nov 24 10:53:00 crc kubenswrapper[4944]: E1124 10:53:00.788180 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18\": container with ID starting with c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18 not found: ID does not exist" containerID="c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.788218 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18"} err="failed to get container status \"c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18\": rpc error: code = NotFound desc = could not find container \"c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18\": container with ID starting with c23a952acc11c6c6f99d1e9578d819845ade93b7912c94dd686b6ac28f9cdd18 not found: ID does not exist" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.788240 4944 scope.go:117] "RemoveContainer" containerID="efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e" Nov 24 10:53:00 crc kubenswrapper[4944]: E1124 10:53:00.788535 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e\": container with ID starting with efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e not found: ID does not exist" containerID="efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.788569 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e"} err="failed to get container status \"efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e\": rpc error: code = NotFound desc = could not find container \"efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e\": container with ID starting with efe8564adf0d60b0125c6c1dfe689e2f8d8199eaf8918ee39a7f9baccc2af65e not found: ID does not exist" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.788589 4944 scope.go:117] "RemoveContainer" containerID="acaa5967079f42908c966a81e230f5dcbf36ca9400dbdaf01f18d598d42350f6" Nov 24 10:53:00 crc kubenswrapper[4944]: E1124 10:53:00.789269 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acaa5967079f42908c966a81e230f5dcbf36ca9400dbdaf01f18d598d42350f6\": container with ID starting with acaa5967079f42908c966a81e230f5dcbf36ca9400dbdaf01f18d598d42350f6 not found: ID does not exist" containerID="acaa5967079f42908c966a81e230f5dcbf36ca9400dbdaf01f18d598d42350f6" Nov 24 10:53:00 crc kubenswrapper[4944]: I1124 10:53:00.789366 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acaa5967079f42908c966a81e230f5dcbf36ca9400dbdaf01f18d598d42350f6"} err="failed to get container status \"acaa5967079f42908c966a81e230f5dcbf36ca9400dbdaf01f18d598d42350f6\": rpc error: code = NotFound desc = could not find container \"acaa5967079f42908c966a81e230f5dcbf36ca9400dbdaf01f18d598d42350f6\": container with ID starting with acaa5967079f42908c966a81e230f5dcbf36ca9400dbdaf01f18d598d42350f6 not found: ID does not exist" Nov 24 10:53:02 crc kubenswrapper[4944]: I1124 10:53:02.289888 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93aeed16-88c6-4027-9efa-94c1286469c5" path="/var/lib/kubelet/pods/93aeed16-88c6-4027-9efa-94c1286469c5/volumes" Nov 24 10:53:06 crc kubenswrapper[4944]: I1124 10:53:06.277273 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:53:06 crc kubenswrapper[4944]: E1124 10:53:06.278614 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:53:12 crc kubenswrapper[4944]: I1124 10:53:12.790209 4944 generic.go:334] "Generic (PLEG): container finished" podID="3317708a-6513-47e2-b70f-b4137e7e25d1" containerID="bbc4f13846541a09bbfd7a55ad449e368f950b468aaaabbf5548e8ea235615a8" exitCode=0 Nov 24 10:53:12 crc kubenswrapper[4944]: I1124 10:53:12.790329 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-89pgd" event={"ID":"3317708a-6513-47e2-b70f-b4137e7e25d1","Type":"ContainerDied","Data":"bbc4f13846541a09bbfd7a55ad449e368f950b468aaaabbf5548e8ea235615a8"} Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.311865 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.455967 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-inventory\") pod \"3317708a-6513-47e2-b70f-b4137e7e25d1\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.456094 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ceph\") pod \"3317708a-6513-47e2-b70f-b4137e7e25d1\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.456316 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fkcf\" (UniqueName: \"kubernetes.io/projected/3317708a-6513-47e2-b70f-b4137e7e25d1-kube-api-access-7fkcf\") pod \"3317708a-6513-47e2-b70f-b4137e7e25d1\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.456377 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ssh-key\") pod \"3317708a-6513-47e2-b70f-b4137e7e25d1\" (UID: \"3317708a-6513-47e2-b70f-b4137e7e25d1\") " Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.461874 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3317708a-6513-47e2-b70f-b4137e7e25d1-kube-api-access-7fkcf" (OuterVolumeSpecName: "kube-api-access-7fkcf") pod "3317708a-6513-47e2-b70f-b4137e7e25d1" (UID: "3317708a-6513-47e2-b70f-b4137e7e25d1"). InnerVolumeSpecName "kube-api-access-7fkcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.461979 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ceph" (OuterVolumeSpecName: "ceph") pod "3317708a-6513-47e2-b70f-b4137e7e25d1" (UID: "3317708a-6513-47e2-b70f-b4137e7e25d1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.494233 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3317708a-6513-47e2-b70f-b4137e7e25d1" (UID: "3317708a-6513-47e2-b70f-b4137e7e25d1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.498222 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-inventory" (OuterVolumeSpecName: "inventory") pod "3317708a-6513-47e2-b70f-b4137e7e25d1" (UID: "3317708a-6513-47e2-b70f-b4137e7e25d1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.559315 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.559782 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.559869 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fkcf\" (UniqueName: \"kubernetes.io/projected/3317708a-6513-47e2-b70f-b4137e7e25d1-kube-api-access-7fkcf\") on node \"crc\" DevicePath \"\"" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.559956 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3317708a-6513-47e2-b70f-b4137e7e25d1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.811090 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-89pgd" event={"ID":"3317708a-6513-47e2-b70f-b4137e7e25d1","Type":"ContainerDied","Data":"0b4fbc4739b629af20f73b3cff18096fc05240dee561e7f1043be841d2d33554"} Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.811156 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b4fbc4739b629af20f73b3cff18096fc05240dee561e7f1043be841d2d33554" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.811231 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-89pgd" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.901713 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-vp6h7"] Nov 24 10:53:14 crc kubenswrapper[4944]: E1124 10:53:14.902586 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93aeed16-88c6-4027-9efa-94c1286469c5" containerName="extract-utilities" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.902612 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="93aeed16-88c6-4027-9efa-94c1286469c5" containerName="extract-utilities" Nov 24 10:53:14 crc kubenswrapper[4944]: E1124 10:53:14.902627 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93aeed16-88c6-4027-9efa-94c1286469c5" containerName="extract-content" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.902634 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="93aeed16-88c6-4027-9efa-94c1286469c5" containerName="extract-content" Nov 24 10:53:14 crc kubenswrapper[4944]: E1124 10:53:14.902651 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3317708a-6513-47e2-b70f-b4137e7e25d1" containerName="install-os-openstack-openstack-cell1" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.902657 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3317708a-6513-47e2-b70f-b4137e7e25d1" containerName="install-os-openstack-openstack-cell1" Nov 24 10:53:14 crc kubenswrapper[4944]: E1124 10:53:14.902686 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" containerName="extract-utilities" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.902749 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" containerName="extract-utilities" Nov 24 10:53:14 crc kubenswrapper[4944]: E1124 10:53:14.902770 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" containerName="registry-server" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.902777 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" containerName="registry-server" Nov 24 10:53:14 crc kubenswrapper[4944]: E1124 10:53:14.902793 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" containerName="extract-content" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.902800 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" containerName="extract-content" Nov 24 10:53:14 crc kubenswrapper[4944]: E1124 10:53:14.902811 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93aeed16-88c6-4027-9efa-94c1286469c5" containerName="registry-server" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.902817 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="93aeed16-88c6-4027-9efa-94c1286469c5" containerName="registry-server" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.903069 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3317708a-6513-47e2-b70f-b4137e7e25d1" containerName="install-os-openstack-openstack-cell1" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.903090 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa6876f3-2ad9-4ac9-ac2a-cac2aef4f667" containerName="registry-server" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.903115 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="93aeed16-88c6-4027-9efa-94c1286469c5" containerName="registry-server" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.905075 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.908927 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.909072 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.908940 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.922041 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-vp6h7"] Nov 24 10:53:14 crc kubenswrapper[4944]: I1124 10:53:14.922750 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.071588 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhklc\" (UniqueName: \"kubernetes.io/projected/9b3afe10-ef54-431f-91b1-b05e0900d544-kube-api-access-fhklc\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.071702 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-inventory\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.071759 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ceph\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.071783 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ssh-key\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.173876 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhklc\" (UniqueName: \"kubernetes.io/projected/9b3afe10-ef54-431f-91b1-b05e0900d544-kube-api-access-fhklc\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.173982 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-inventory\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.174096 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ceph\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.174130 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ssh-key\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.177826 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ssh-key\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.178493 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-inventory\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.178638 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ceph\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.198404 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhklc\" (UniqueName: \"kubernetes.io/projected/9b3afe10-ef54-431f-91b1-b05e0900d544-kube-api-access-fhklc\") pod \"configure-os-openstack-openstack-cell1-vp6h7\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.235874 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.573514 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-vp6h7"] Nov 24 10:53:15 crc kubenswrapper[4944]: I1124 10:53:15.819969 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" event={"ID":"9b3afe10-ef54-431f-91b1-b05e0900d544","Type":"ContainerStarted","Data":"d038fce66cc116f3c534541810e2630c80f5faa94c1949342d1f102a12c5437c"} Nov 24 10:53:16 crc kubenswrapper[4944]: I1124 10:53:16.828149 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" event={"ID":"9b3afe10-ef54-431f-91b1-b05e0900d544","Type":"ContainerStarted","Data":"65aee8fd6759e527f403db3d6e47cbe556ea2eea1e782e615829eff20cad992b"} Nov 24 10:53:16 crc kubenswrapper[4944]: I1124 10:53:16.849441 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" podStartSLOduration=2.399045818 podStartE2EDuration="2.84942468s" podCreationTimestamp="2025-11-24 10:53:14 +0000 UTC" firstStartedPulling="2025-11-24 10:53:15.57962943 +0000 UTC m=+7256.114069892" lastFinishedPulling="2025-11-24 10:53:16.030008292 +0000 UTC m=+7256.564448754" observedRunningTime="2025-11-24 10:53:16.841001332 +0000 UTC m=+7257.375441794" watchObservedRunningTime="2025-11-24 10:53:16.84942468 +0000 UTC m=+7257.383865142" Nov 24 10:53:18 crc kubenswrapper[4944]: I1124 10:53:18.277704 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:53:18 crc kubenswrapper[4944]: E1124 10:53:18.278592 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:53:30 crc kubenswrapper[4944]: I1124 10:53:30.286392 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:53:30 crc kubenswrapper[4944]: E1124 10:53:30.287475 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:53:41 crc kubenswrapper[4944]: I1124 10:53:41.277407 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:53:41 crc kubenswrapper[4944]: E1124 10:53:41.278713 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:53:54 crc kubenswrapper[4944]: I1124 10:53:54.280642 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:53:54 crc kubenswrapper[4944]: E1124 10:53:54.281661 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:53:59 crc kubenswrapper[4944]: I1124 10:53:59.323386 4944 generic.go:334] "Generic (PLEG): container finished" podID="9b3afe10-ef54-431f-91b1-b05e0900d544" containerID="65aee8fd6759e527f403db3d6e47cbe556ea2eea1e782e615829eff20cad992b" exitCode=0 Nov 24 10:53:59 crc kubenswrapper[4944]: I1124 10:53:59.323501 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" event={"ID":"9b3afe10-ef54-431f-91b1-b05e0900d544","Type":"ContainerDied","Data":"65aee8fd6759e527f403db3d6e47cbe556ea2eea1e782e615829eff20cad992b"} Nov 24 10:54:00 crc kubenswrapper[4944]: I1124 10:54:00.829410 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:54:00 crc kubenswrapper[4944]: I1124 10:54:00.984034 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ceph\") pod \"9b3afe10-ef54-431f-91b1-b05e0900d544\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " Nov 24 10:54:00 crc kubenswrapper[4944]: I1124 10:54:00.984162 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ssh-key\") pod \"9b3afe10-ef54-431f-91b1-b05e0900d544\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " Nov 24 10:54:00 crc kubenswrapper[4944]: I1124 10:54:00.984261 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhklc\" (UniqueName: \"kubernetes.io/projected/9b3afe10-ef54-431f-91b1-b05e0900d544-kube-api-access-fhklc\") pod \"9b3afe10-ef54-431f-91b1-b05e0900d544\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " Nov 24 10:54:00 crc kubenswrapper[4944]: I1124 10:54:00.984408 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-inventory\") pod \"9b3afe10-ef54-431f-91b1-b05e0900d544\" (UID: \"9b3afe10-ef54-431f-91b1-b05e0900d544\") " Nov 24 10:54:00 crc kubenswrapper[4944]: I1124 10:54:00.994356 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ceph" (OuterVolumeSpecName: "ceph") pod "9b3afe10-ef54-431f-91b1-b05e0900d544" (UID: "9b3afe10-ef54-431f-91b1-b05e0900d544"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:00 crc kubenswrapper[4944]: I1124 10:54:00.998545 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b3afe10-ef54-431f-91b1-b05e0900d544-kube-api-access-fhklc" (OuterVolumeSpecName: "kube-api-access-fhklc") pod "9b3afe10-ef54-431f-91b1-b05e0900d544" (UID: "9b3afe10-ef54-431f-91b1-b05e0900d544"). InnerVolumeSpecName "kube-api-access-fhklc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.027244 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-inventory" (OuterVolumeSpecName: "inventory") pod "9b3afe10-ef54-431f-91b1-b05e0900d544" (UID: "9b3afe10-ef54-431f-91b1-b05e0900d544"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.028428 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9b3afe10-ef54-431f-91b1-b05e0900d544" (UID: "9b3afe10-ef54-431f-91b1-b05e0900d544"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.087253 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.087288 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.087298 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b3afe10-ef54-431f-91b1-b05e0900d544-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.087307 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhklc\" (UniqueName: \"kubernetes.io/projected/9b3afe10-ef54-431f-91b1-b05e0900d544-kube-api-access-fhklc\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.346909 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" event={"ID":"9b3afe10-ef54-431f-91b1-b05e0900d544","Type":"ContainerDied","Data":"d038fce66cc116f3c534541810e2630c80f5faa94c1949342d1f102a12c5437c"} Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.346952 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d038fce66cc116f3c534541810e2630c80f5faa94c1949342d1f102a12c5437c" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.347005 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-vp6h7" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.423865 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-h68cc"] Nov 24 10:54:01 crc kubenswrapper[4944]: E1124 10:54:01.425644 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b3afe10-ef54-431f-91b1-b05e0900d544" containerName="configure-os-openstack-openstack-cell1" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.425749 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b3afe10-ef54-431f-91b1-b05e0900d544" containerName="configure-os-openstack-openstack-cell1" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.426193 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b3afe10-ef54-431f-91b1-b05e0900d544" containerName="configure-os-openstack-openstack-cell1" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.432248 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.435536 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-h68cc"] Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.441584 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.441761 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.442042 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.442410 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.598008 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.598103 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ceph\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.598137 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d44rg\" (UniqueName: \"kubernetes.io/projected/e3efa51f-a625-4117-85ff-1eb6f5016b3a-kube-api-access-d44rg\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.598312 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-inventory-0\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.701416 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ceph\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.701492 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d44rg\" (UniqueName: \"kubernetes.io/projected/e3efa51f-a625-4117-85ff-1eb6f5016b3a-kube-api-access-d44rg\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.701543 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-inventory-0\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.701744 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.706735 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-inventory-0\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.706740 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ceph\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.710573 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.719946 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d44rg\" (UniqueName: \"kubernetes.io/projected/e3efa51f-a625-4117-85ff-1eb6f5016b3a-kube-api-access-d44rg\") pod \"ssh-known-hosts-openstack-h68cc\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:01 crc kubenswrapper[4944]: I1124 10:54:01.756351 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:02 crc kubenswrapper[4944]: I1124 10:54:02.300011 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-h68cc"] Nov 24 10:54:02 crc kubenswrapper[4944]: W1124 10:54:02.307527 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3efa51f_a625_4117_85ff_1eb6f5016b3a.slice/crio-580ee00bf252fd9a1a111e4303424f269f3edc797204fb3926a3bba28464ac59 WatchSource:0}: Error finding container 580ee00bf252fd9a1a111e4303424f269f3edc797204fb3926a3bba28464ac59: Status 404 returned error can't find the container with id 580ee00bf252fd9a1a111e4303424f269f3edc797204fb3926a3bba28464ac59 Nov 24 10:54:02 crc kubenswrapper[4944]: I1124 10:54:02.356234 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h68cc" event={"ID":"e3efa51f-a625-4117-85ff-1eb6f5016b3a","Type":"ContainerStarted","Data":"580ee00bf252fd9a1a111e4303424f269f3edc797204fb3926a3bba28464ac59"} Nov 24 10:54:03 crc kubenswrapper[4944]: I1124 10:54:03.366874 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h68cc" event={"ID":"e3efa51f-a625-4117-85ff-1eb6f5016b3a","Type":"ContainerStarted","Data":"91328e0c57780a8bf7b41dfb6d474120c3a5c9a727f83abd9d700055ac46329c"} Nov 24 10:54:03 crc kubenswrapper[4944]: I1124 10:54:03.383570 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-h68cc" podStartSLOduration=2.001110864 podStartE2EDuration="2.383552416s" podCreationTimestamp="2025-11-24 10:54:01 +0000 UTC" firstStartedPulling="2025-11-24 10:54:02.309397625 +0000 UTC m=+7302.843838097" lastFinishedPulling="2025-11-24 10:54:02.691839147 +0000 UTC m=+7303.226279649" observedRunningTime="2025-11-24 10:54:03.381558743 +0000 UTC m=+7303.915999205" watchObservedRunningTime="2025-11-24 10:54:03.383552416 +0000 UTC m=+7303.917992878" Nov 24 10:54:06 crc kubenswrapper[4944]: I1124 10:54:06.278318 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:54:06 crc kubenswrapper[4944]: E1124 10:54:06.279412 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:54:10 crc kubenswrapper[4944]: I1124 10:54:10.433344 4944 generic.go:334] "Generic (PLEG): container finished" podID="e3efa51f-a625-4117-85ff-1eb6f5016b3a" containerID="91328e0c57780a8bf7b41dfb6d474120c3a5c9a727f83abd9d700055ac46329c" exitCode=0 Nov 24 10:54:10 crc kubenswrapper[4944]: I1124 10:54:10.433444 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h68cc" event={"ID":"e3efa51f-a625-4117-85ff-1eb6f5016b3a","Type":"ContainerDied","Data":"91328e0c57780a8bf7b41dfb6d474120c3a5c9a727f83abd9d700055ac46329c"} Nov 24 10:54:11 crc kubenswrapper[4944]: I1124 10:54:11.930765 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.017867 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ssh-key-openstack-cell1\") pod \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.018011 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ceph\") pod \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.018077 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-inventory-0\") pod \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.018125 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d44rg\" (UniqueName: \"kubernetes.io/projected/e3efa51f-a625-4117-85ff-1eb6f5016b3a-kube-api-access-d44rg\") pod \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\" (UID: \"e3efa51f-a625-4117-85ff-1eb6f5016b3a\") " Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.023167 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ceph" (OuterVolumeSpecName: "ceph") pod "e3efa51f-a625-4117-85ff-1eb6f5016b3a" (UID: "e3efa51f-a625-4117-85ff-1eb6f5016b3a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.023247 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3efa51f-a625-4117-85ff-1eb6f5016b3a-kube-api-access-d44rg" (OuterVolumeSpecName: "kube-api-access-d44rg") pod "e3efa51f-a625-4117-85ff-1eb6f5016b3a" (UID: "e3efa51f-a625-4117-85ff-1eb6f5016b3a"). InnerVolumeSpecName "kube-api-access-d44rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.045555 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "e3efa51f-a625-4117-85ff-1eb6f5016b3a" (UID: "e3efa51f-a625-4117-85ff-1eb6f5016b3a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.046585 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "e3efa51f-a625-4117-85ff-1eb6f5016b3a" (UID: "e3efa51f-a625-4117-85ff-1eb6f5016b3a"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.123258 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.123378 4944 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.123507 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d44rg\" (UniqueName: \"kubernetes.io/projected/e3efa51f-a625-4117-85ff-1eb6f5016b3a-kube-api-access-d44rg\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.123591 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/e3efa51f-a625-4117-85ff-1eb6f5016b3a-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.456679 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h68cc" event={"ID":"e3efa51f-a625-4117-85ff-1eb6f5016b3a","Type":"ContainerDied","Data":"580ee00bf252fd9a1a111e4303424f269f3edc797204fb3926a3bba28464ac59"} Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.456971 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="580ee00bf252fd9a1a111e4303424f269f3edc797204fb3926a3bba28464ac59" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.456749 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h68cc" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.531648 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-tdz6j"] Nov 24 10:54:12 crc kubenswrapper[4944]: E1124 10:54:12.532572 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3efa51f-a625-4117-85ff-1eb6f5016b3a" containerName="ssh-known-hosts-openstack" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.532594 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3efa51f-a625-4117-85ff-1eb6f5016b3a" containerName="ssh-known-hosts-openstack" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.532925 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3efa51f-a625-4117-85ff-1eb6f5016b3a" containerName="ssh-known-hosts-openstack" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.534196 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.537084 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.537188 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.537265 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.537462 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.540731 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-tdz6j"] Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.634753 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ceph\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.634830 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ssh-key\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.634878 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwwh9\" (UniqueName: \"kubernetes.io/projected/85c42bbf-4cef-46a7-9935-108aabab6020-kube-api-access-cwwh9\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.634926 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-inventory\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.736895 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ceph\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.736959 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ssh-key\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.736994 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwwh9\" (UniqueName: \"kubernetes.io/projected/85c42bbf-4cef-46a7-9935-108aabab6020-kube-api-access-cwwh9\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.737036 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-inventory\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.740376 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ssh-key\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.740496 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-inventory\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.745502 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ceph\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.752985 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwwh9\" (UniqueName: \"kubernetes.io/projected/85c42bbf-4cef-46a7-9935-108aabab6020-kube-api-access-cwwh9\") pod \"run-os-openstack-openstack-cell1-tdz6j\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:12 crc kubenswrapper[4944]: I1124 10:54:12.865291 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:13 crc kubenswrapper[4944]: I1124 10:54:13.373263 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-tdz6j"] Nov 24 10:54:13 crc kubenswrapper[4944]: I1124 10:54:13.466941 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tdz6j" event={"ID":"85c42bbf-4cef-46a7-9935-108aabab6020","Type":"ContainerStarted","Data":"61fbe4d993f2ccd2a79b4204d32c9005887a348a543580998587a5d59cd719ab"} Nov 24 10:54:14 crc kubenswrapper[4944]: I1124 10:54:14.479987 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tdz6j" event={"ID":"85c42bbf-4cef-46a7-9935-108aabab6020","Type":"ContainerStarted","Data":"4de2400b470003c80a575d5c50d4bc141102bbec49a605446f836efa639430ae"} Nov 24 10:54:14 crc kubenswrapper[4944]: I1124 10:54:14.503478 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-tdz6j" podStartSLOduration=2.094171363 podStartE2EDuration="2.503460519s" podCreationTimestamp="2025-11-24 10:54:12 +0000 UTC" firstStartedPulling="2025-11-24 10:54:13.383132 +0000 UTC m=+7313.917572452" lastFinishedPulling="2025-11-24 10:54:13.792421136 +0000 UTC m=+7314.326861608" observedRunningTime="2025-11-24 10:54:14.492535892 +0000 UTC m=+7315.026976374" watchObservedRunningTime="2025-11-24 10:54:14.503460519 +0000 UTC m=+7315.037900971" Nov 24 10:54:18 crc kubenswrapper[4944]: I1124 10:54:18.278028 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:54:18 crc kubenswrapper[4944]: E1124 10:54:18.278696 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:54:21 crc kubenswrapper[4944]: I1124 10:54:21.547709 4944 generic.go:334] "Generic (PLEG): container finished" podID="85c42bbf-4cef-46a7-9935-108aabab6020" containerID="4de2400b470003c80a575d5c50d4bc141102bbec49a605446f836efa639430ae" exitCode=0 Nov 24 10:54:21 crc kubenswrapper[4944]: I1124 10:54:21.547844 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tdz6j" event={"ID":"85c42bbf-4cef-46a7-9935-108aabab6020","Type":"ContainerDied","Data":"4de2400b470003c80a575d5c50d4bc141102bbec49a605446f836efa639430ae"} Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.110344 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.164861 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-inventory\") pod \"85c42bbf-4cef-46a7-9935-108aabab6020\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.164988 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ceph\") pod \"85c42bbf-4cef-46a7-9935-108aabab6020\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.165043 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwwh9\" (UniqueName: \"kubernetes.io/projected/85c42bbf-4cef-46a7-9935-108aabab6020-kube-api-access-cwwh9\") pod \"85c42bbf-4cef-46a7-9935-108aabab6020\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.165243 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ssh-key\") pod \"85c42bbf-4cef-46a7-9935-108aabab6020\" (UID: \"85c42bbf-4cef-46a7-9935-108aabab6020\") " Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.171006 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ceph" (OuterVolumeSpecName: "ceph") pod "85c42bbf-4cef-46a7-9935-108aabab6020" (UID: "85c42bbf-4cef-46a7-9935-108aabab6020"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.171355 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c42bbf-4cef-46a7-9935-108aabab6020-kube-api-access-cwwh9" (OuterVolumeSpecName: "kube-api-access-cwwh9") pod "85c42bbf-4cef-46a7-9935-108aabab6020" (UID: "85c42bbf-4cef-46a7-9935-108aabab6020"). InnerVolumeSpecName "kube-api-access-cwwh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.197325 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-inventory" (OuterVolumeSpecName: "inventory") pod "85c42bbf-4cef-46a7-9935-108aabab6020" (UID: "85c42bbf-4cef-46a7-9935-108aabab6020"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.203975 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "85c42bbf-4cef-46a7-9935-108aabab6020" (UID: "85c42bbf-4cef-46a7-9935-108aabab6020"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.268016 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.268088 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwwh9\" (UniqueName: \"kubernetes.io/projected/85c42bbf-4cef-46a7-9935-108aabab6020-kube-api-access-cwwh9\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.268102 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.268115 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85c42bbf-4cef-46a7-9935-108aabab6020-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.576729 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tdz6j" event={"ID":"85c42bbf-4cef-46a7-9935-108aabab6020","Type":"ContainerDied","Data":"61fbe4d993f2ccd2a79b4204d32c9005887a348a543580998587a5d59cd719ab"} Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.577519 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61fbe4d993f2ccd2a79b4204d32c9005887a348a543580998587a5d59cd719ab" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.576802 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tdz6j" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.635719 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-c8btv"] Nov 24 10:54:23 crc kubenswrapper[4944]: E1124 10:54:23.636474 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c42bbf-4cef-46a7-9935-108aabab6020" containerName="run-os-openstack-openstack-cell1" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.636610 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c42bbf-4cef-46a7-9935-108aabab6020" containerName="run-os-openstack-openstack-cell1" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.637008 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c42bbf-4cef-46a7-9935-108aabab6020" containerName="run-os-openstack-openstack-cell1" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.638118 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.641152 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.641853 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.642194 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.642534 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.648212 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-c8btv"] Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.678447 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.678539 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-inventory\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.678582 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2snqr\" (UniqueName: \"kubernetes.io/projected/f69ee93c-d8e0-4c43-8597-21050c074c71-kube-api-access-2snqr\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.678738 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ceph\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.780185 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.780280 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-inventory\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.780325 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2snqr\" (UniqueName: \"kubernetes.io/projected/f69ee93c-d8e0-4c43-8597-21050c074c71-kube-api-access-2snqr\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.780522 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ceph\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.786269 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ceph\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.790015 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-inventory\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.797698 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.809867 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2snqr\" (UniqueName: \"kubernetes.io/projected/f69ee93c-d8e0-4c43-8597-21050c074c71-kube-api-access-2snqr\") pod \"reboot-os-openstack-openstack-cell1-c8btv\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:23 crc kubenswrapper[4944]: I1124 10:54:23.961648 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:24 crc kubenswrapper[4944]: I1124 10:54:24.451971 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-c8btv"] Nov 24 10:54:24 crc kubenswrapper[4944]: I1124 10:54:24.458662 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 10:54:24 crc kubenswrapper[4944]: I1124 10:54:24.585809 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" event={"ID":"f69ee93c-d8e0-4c43-8597-21050c074c71","Type":"ContainerStarted","Data":"e52bceee02e721d1b58e59caa27782b366f3e1ee6d6147663343650f15bd2474"} Nov 24 10:54:25 crc kubenswrapper[4944]: I1124 10:54:25.594308 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" event={"ID":"f69ee93c-d8e0-4c43-8597-21050c074c71","Type":"ContainerStarted","Data":"23de47c3b81d06adf7895ef8ce4422df4f48cf2ca4b38dec73db2013e6931a15"} Nov 24 10:54:25 crc kubenswrapper[4944]: I1124 10:54:25.627702 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" podStartSLOduration=2.101895793 podStartE2EDuration="2.62767624s" podCreationTimestamp="2025-11-24 10:54:23 +0000 UTC" firstStartedPulling="2025-11-24 10:54:24.458465037 +0000 UTC m=+7324.992905499" lastFinishedPulling="2025-11-24 10:54:24.984245484 +0000 UTC m=+7325.518685946" observedRunningTime="2025-11-24 10:54:25.610466874 +0000 UTC m=+7326.144907336" watchObservedRunningTime="2025-11-24 10:54:25.62767624 +0000 UTC m=+7326.162116732" Nov 24 10:54:31 crc kubenswrapper[4944]: I1124 10:54:31.277570 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:54:31 crc kubenswrapper[4944]: E1124 10:54:31.278406 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:54:41 crc kubenswrapper[4944]: I1124 10:54:41.766240 4944 generic.go:334] "Generic (PLEG): container finished" podID="f69ee93c-d8e0-4c43-8597-21050c074c71" containerID="23de47c3b81d06adf7895ef8ce4422df4f48cf2ca4b38dec73db2013e6931a15" exitCode=0 Nov 24 10:54:41 crc kubenswrapper[4944]: I1124 10:54:41.766338 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" event={"ID":"f69ee93c-d8e0-4c43-8597-21050c074c71","Type":"ContainerDied","Data":"23de47c3b81d06adf7895ef8ce4422df4f48cf2ca4b38dec73db2013e6931a15"} Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.285815 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.368391 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-inventory\") pod \"f69ee93c-d8e0-4c43-8597-21050c074c71\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.368448 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ssh-key\") pod \"f69ee93c-d8e0-4c43-8597-21050c074c71\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.368665 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ceph\") pod \"f69ee93c-d8e0-4c43-8597-21050c074c71\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.368751 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2snqr\" (UniqueName: \"kubernetes.io/projected/f69ee93c-d8e0-4c43-8597-21050c074c71-kube-api-access-2snqr\") pod \"f69ee93c-d8e0-4c43-8597-21050c074c71\" (UID: \"f69ee93c-d8e0-4c43-8597-21050c074c71\") " Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.374214 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ceph" (OuterVolumeSpecName: "ceph") pod "f69ee93c-d8e0-4c43-8597-21050c074c71" (UID: "f69ee93c-d8e0-4c43-8597-21050c074c71"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.376358 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69ee93c-d8e0-4c43-8597-21050c074c71-kube-api-access-2snqr" (OuterVolumeSpecName: "kube-api-access-2snqr") pod "f69ee93c-d8e0-4c43-8597-21050c074c71" (UID: "f69ee93c-d8e0-4c43-8597-21050c074c71"). InnerVolumeSpecName "kube-api-access-2snqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.405239 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-inventory" (OuterVolumeSpecName: "inventory") pod "f69ee93c-d8e0-4c43-8597-21050c074c71" (UID: "f69ee93c-d8e0-4c43-8597-21050c074c71"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.405460 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f69ee93c-d8e0-4c43-8597-21050c074c71" (UID: "f69ee93c-d8e0-4c43-8597-21050c074c71"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.472719 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2snqr\" (UniqueName: \"kubernetes.io/projected/f69ee93c-d8e0-4c43-8597-21050c074c71-kube-api-access-2snqr\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.472760 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.472772 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.472787 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f69ee93c-d8e0-4c43-8597-21050c074c71-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.785981 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" event={"ID":"f69ee93c-d8e0-4c43-8597-21050c074c71","Type":"ContainerDied","Data":"e52bceee02e721d1b58e59caa27782b366f3e1ee6d6147663343650f15bd2474"} Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.786302 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e52bceee02e721d1b58e59caa27782b366f3e1ee6d6147663343650f15bd2474" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.786026 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-c8btv" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.867930 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-gw8rw"] Nov 24 10:54:43 crc kubenswrapper[4944]: E1124 10:54:43.868426 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69ee93c-d8e0-4c43-8597-21050c074c71" containerName="reboot-os-openstack-openstack-cell1" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.868444 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69ee93c-d8e0-4c43-8597-21050c074c71" containerName="reboot-os-openstack-openstack-cell1" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.868669 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f69ee93c-d8e0-4c43-8597-21050c074c71" containerName="reboot-os-openstack-openstack-cell1" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.869413 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.871474 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.871964 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.872172 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.873264 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.880434 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-gw8rw"] Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.983394 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.983479 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.983514 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-inventory\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.983642 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ssh-key\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.983713 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.983764 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.983887 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.983950 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.984033 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.984174 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x2fz\" (UniqueName: \"kubernetes.io/projected/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-kube-api-access-9x2fz\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.984208 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:43 crc kubenswrapper[4944]: I1124 10:54:43.984248 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ceph\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.086024 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ssh-key\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.087134 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.087208 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.087242 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.087301 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.087368 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.087483 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x2fz\" (UniqueName: \"kubernetes.io/projected/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-kube-api-access-9x2fz\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.087515 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.087564 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ceph\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.087606 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.087655 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.087683 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-inventory\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.093915 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ceph\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.094044 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.094319 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ssh-key\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.093226 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.097418 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-inventory\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.097879 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.098222 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.098825 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.099791 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.109747 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.114569 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x2fz\" (UniqueName: \"kubernetes.io/projected/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-kube-api-access-9x2fz\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.123020 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-gw8rw\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.187815 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.276508 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:54:44 crc kubenswrapper[4944]: E1124 10:54:44.277156 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 10:54:44 crc kubenswrapper[4944]: W1124 10:54:44.748361 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbde6f616_2438_4d9d_8c87_a0f8257ab2fb.slice/crio-f101ad58dd7fbc1ba5bd75b5d60b8c14e958ec82cdfffbe9b3baa7fd94c3bf33 WatchSource:0}: Error finding container f101ad58dd7fbc1ba5bd75b5d60b8c14e958ec82cdfffbe9b3baa7fd94c3bf33: Status 404 returned error can't find the container with id f101ad58dd7fbc1ba5bd75b5d60b8c14e958ec82cdfffbe9b3baa7fd94c3bf33 Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.748605 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-gw8rw"] Nov 24 10:54:44 crc kubenswrapper[4944]: I1124 10:54:44.797258 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" event={"ID":"bde6f616-2438-4d9d-8c87-a0f8257ab2fb","Type":"ContainerStarted","Data":"f101ad58dd7fbc1ba5bd75b5d60b8c14e958ec82cdfffbe9b3baa7fd94c3bf33"} Nov 24 10:54:45 crc kubenswrapper[4944]: I1124 10:54:45.807357 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" event={"ID":"bde6f616-2438-4d9d-8c87-a0f8257ab2fb","Type":"ContainerStarted","Data":"c193b4b6f77fd276635d105fbb0711f9bb1b3899ba945400c859286f95957ed7"} Nov 24 10:54:45 crc kubenswrapper[4944]: I1124 10:54:45.832030 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" podStartSLOduration=2.3275147880000002 podStartE2EDuration="2.832008699s" podCreationTimestamp="2025-11-24 10:54:43 +0000 UTC" firstStartedPulling="2025-11-24 10:54:44.751606028 +0000 UTC m=+7345.286046520" lastFinishedPulling="2025-11-24 10:54:45.256099969 +0000 UTC m=+7345.790540431" observedRunningTime="2025-11-24 10:54:45.826112172 +0000 UTC m=+7346.360552634" watchObservedRunningTime="2025-11-24 10:54:45.832008699 +0000 UTC m=+7346.366449161" Nov 24 10:54:56 crc kubenswrapper[4944]: I1124 10:54:56.277662 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:54:56 crc kubenswrapper[4944]: I1124 10:54:56.906500 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"711f3ecca590ed13334ef816a5b7121fb9369df33f5c8c0821dcbd83e7e9d442"} Nov 24 10:55:03 crc kubenswrapper[4944]: I1124 10:55:03.978356 4944 generic.go:334] "Generic (PLEG): container finished" podID="bde6f616-2438-4d9d-8c87-a0f8257ab2fb" containerID="c193b4b6f77fd276635d105fbb0711f9bb1b3899ba945400c859286f95957ed7" exitCode=0 Nov 24 10:55:03 crc kubenswrapper[4944]: I1124 10:55:03.978447 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" event={"ID":"bde6f616-2438-4d9d-8c87-a0f8257ab2fb","Type":"ContainerDied","Data":"c193b4b6f77fd276635d105fbb0711f9bb1b3899ba945400c859286f95957ed7"} Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.493398 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563306 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-metadata-combined-ca-bundle\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563346 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-dhcp-combined-ca-bundle\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563416 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-inventory\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563440 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-sriov-combined-ca-bundle\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563464 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x2fz\" (UniqueName: \"kubernetes.io/projected/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-kube-api-access-9x2fz\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563495 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-libvirt-combined-ca-bundle\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563548 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-bootstrap-combined-ca-bundle\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563573 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ssh-key\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563651 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ceph\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563682 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-nova-combined-ca-bundle\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563707 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-telemetry-combined-ca-bundle\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.563736 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ovn-combined-ca-bundle\") pod \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\" (UID: \"bde6f616-2438-4d9d-8c87-a0f8257ab2fb\") " Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.569344 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.569399 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.570863 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.572815 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.573089 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.573386 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ceph" (OuterVolumeSpecName: "ceph") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.573423 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.574146 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-kube-api-access-9x2fz" (OuterVolumeSpecName: "kube-api-access-9x2fz") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "kube-api-access-9x2fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.574240 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.574567 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.597091 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.600232 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-inventory" (OuterVolumeSpecName: "inventory") pod "bde6f616-2438-4d9d-8c87-a0f8257ab2fb" (UID: "bde6f616-2438-4d9d-8c87-a0f8257ab2fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666065 4944 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666160 4944 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666201 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666215 4944 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666229 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x2fz\" (UniqueName: \"kubernetes.io/projected/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-kube-api-access-9x2fz\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666241 4944 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666253 4944 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666264 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666276 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666286 4944 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666298 4944 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.666310 4944 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bde6f616-2438-4d9d-8c87-a0f8257ab2fb-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.996315 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" event={"ID":"bde6f616-2438-4d9d-8c87-a0f8257ab2fb","Type":"ContainerDied","Data":"f101ad58dd7fbc1ba5bd75b5d60b8c14e958ec82cdfffbe9b3baa7fd94c3bf33"} Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.996562 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f101ad58dd7fbc1ba5bd75b5d60b8c14e958ec82cdfffbe9b3baa7fd94c3bf33" Nov 24 10:55:05 crc kubenswrapper[4944]: I1124 10:55:05.996372 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-gw8rw" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.147453 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-mw6c6"] Nov 24 10:55:06 crc kubenswrapper[4944]: E1124 10:55:06.148191 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bde6f616-2438-4d9d-8c87-a0f8257ab2fb" containerName="install-certs-openstack-openstack-cell1" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.148283 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bde6f616-2438-4d9d-8c87-a0f8257ab2fb" containerName="install-certs-openstack-openstack-cell1" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.148547 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bde6f616-2438-4d9d-8c87-a0f8257ab2fb" containerName="install-certs-openstack-openstack-cell1" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.149398 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.152637 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.153040 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.153297 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.153613 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.156643 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-mw6c6"] Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.290493 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrz9s\" (UniqueName: \"kubernetes.io/projected/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-kube-api-access-wrz9s\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.290744 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ceph\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.290794 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.290870 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-inventory\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.393009 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ceph\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.393122 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.393251 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-inventory\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.393495 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrz9s\" (UniqueName: \"kubernetes.io/projected/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-kube-api-access-wrz9s\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.400821 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ceph\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.403550 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.410266 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-inventory\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.416227 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrz9s\" (UniqueName: \"kubernetes.io/projected/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-kube-api-access-wrz9s\") pod \"ceph-client-openstack-openstack-cell1-mw6c6\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:06 crc kubenswrapper[4944]: I1124 10:55:06.484962 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:07 crc kubenswrapper[4944]: I1124 10:55:07.029471 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-mw6c6"] Nov 24 10:55:08 crc kubenswrapper[4944]: I1124 10:55:08.021775 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" event={"ID":"23ab44ae-eb0d-47a7-b321-7b53ab21ddff","Type":"ContainerStarted","Data":"209966ba79c77d2145475f2df1cbc46fdba3312797e28fde4b58717589013c2d"} Nov 24 10:55:08 crc kubenswrapper[4944]: I1124 10:55:08.022219 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" event={"ID":"23ab44ae-eb0d-47a7-b321-7b53ab21ddff","Type":"ContainerStarted","Data":"aae16aef2f33adcf1f5e47d40dc615332f8fc18b4a4cffe664da7e2a3f2248de"} Nov 24 10:55:13 crc kubenswrapper[4944]: I1124 10:55:13.071899 4944 generic.go:334] "Generic (PLEG): container finished" podID="23ab44ae-eb0d-47a7-b321-7b53ab21ddff" containerID="209966ba79c77d2145475f2df1cbc46fdba3312797e28fde4b58717589013c2d" exitCode=0 Nov 24 10:55:13 crc kubenswrapper[4944]: I1124 10:55:13.071992 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" event={"ID":"23ab44ae-eb0d-47a7-b321-7b53ab21ddff","Type":"ContainerDied","Data":"209966ba79c77d2145475f2df1cbc46fdba3312797e28fde4b58717589013c2d"} Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.561662 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.680345 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrz9s\" (UniqueName: \"kubernetes.io/projected/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-kube-api-access-wrz9s\") pod \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.680451 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-inventory\") pod \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.680484 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ceph\") pod \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.680678 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ssh-key\") pod \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\" (UID: \"23ab44ae-eb0d-47a7-b321-7b53ab21ddff\") " Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.687209 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ceph" (OuterVolumeSpecName: "ceph") pod "23ab44ae-eb0d-47a7-b321-7b53ab21ddff" (UID: "23ab44ae-eb0d-47a7-b321-7b53ab21ddff"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.687630 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-kube-api-access-wrz9s" (OuterVolumeSpecName: "kube-api-access-wrz9s") pod "23ab44ae-eb0d-47a7-b321-7b53ab21ddff" (UID: "23ab44ae-eb0d-47a7-b321-7b53ab21ddff"). InnerVolumeSpecName "kube-api-access-wrz9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.711040 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-inventory" (OuterVolumeSpecName: "inventory") pod "23ab44ae-eb0d-47a7-b321-7b53ab21ddff" (UID: "23ab44ae-eb0d-47a7-b321-7b53ab21ddff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.712555 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "23ab44ae-eb0d-47a7-b321-7b53ab21ddff" (UID: "23ab44ae-eb0d-47a7-b321-7b53ab21ddff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.782537 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.782566 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.782575 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:14 crc kubenswrapper[4944]: I1124 10:55:14.782583 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrz9s\" (UniqueName: \"kubernetes.io/projected/23ab44ae-eb0d-47a7-b321-7b53ab21ddff-kube-api-access-wrz9s\") on node \"crc\" DevicePath \"\"" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.091650 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" event={"ID":"23ab44ae-eb0d-47a7-b321-7b53ab21ddff","Type":"ContainerDied","Data":"aae16aef2f33adcf1f5e47d40dc615332f8fc18b4a4cffe664da7e2a3f2248de"} Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.091686 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aae16aef2f33adcf1f5e47d40dc615332f8fc18b4a4cffe664da7e2a3f2248de" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.091743 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-mw6c6" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.186068 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-xmptp"] Nov 24 10:55:15 crc kubenswrapper[4944]: E1124 10:55:15.186533 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ab44ae-eb0d-47a7-b321-7b53ab21ddff" containerName="ceph-client-openstack-openstack-cell1" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.186547 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ab44ae-eb0d-47a7-b321-7b53ab21ddff" containerName="ceph-client-openstack-openstack-cell1" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.186724 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ab44ae-eb0d-47a7-b321-7b53ab21ddff" containerName="ceph-client-openstack-openstack-cell1" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.187487 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.191088 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.191296 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.191388 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.191404 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.191522 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.210649 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-xmptp"] Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.294940 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-inventory\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.295328 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t8zt\" (UniqueName: \"kubernetes.io/projected/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-kube-api-access-8t8zt\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.295378 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ssh-key\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.295548 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.295757 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ceph\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.295881 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.397980 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-inventory\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.398070 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t8zt\" (UniqueName: \"kubernetes.io/projected/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-kube-api-access-8t8zt\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.398108 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ssh-key\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.398162 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.398234 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ceph\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.398277 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.399937 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.403987 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-inventory\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.404163 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ssh-key\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.412528 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.416234 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ceph\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.417450 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t8zt\" (UniqueName: \"kubernetes.io/projected/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-kube-api-access-8t8zt\") pod \"ovn-openstack-openstack-cell1-xmptp\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:15 crc kubenswrapper[4944]: I1124 10:55:15.514076 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:55:16 crc kubenswrapper[4944]: I1124 10:55:16.045448 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-xmptp"] Nov 24 10:55:16 crc kubenswrapper[4944]: I1124 10:55:16.102534 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-xmptp" event={"ID":"1362e83f-d5d5-43c2-bba1-21fa8ec0833e","Type":"ContainerStarted","Data":"e15916b3fd2e63ca6530ca214da5d8df5dd6429333a8253a7885f30ef038c535"} Nov 24 10:55:17 crc kubenswrapper[4944]: I1124 10:55:17.113124 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-xmptp" event={"ID":"1362e83f-d5d5-43c2-bba1-21fa8ec0833e","Type":"ContainerStarted","Data":"74c022fe962256cd61e8018ba3318d965a79ac9eb2e6d6d7a20a43a4b005ec50"} Nov 24 10:55:17 crc kubenswrapper[4944]: I1124 10:55:17.137943 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-xmptp" podStartSLOduration=1.727088042 podStartE2EDuration="2.137926627s" podCreationTimestamp="2025-11-24 10:55:15 +0000 UTC" firstStartedPulling="2025-11-24 10:55:16.057153384 +0000 UTC m=+7376.591593836" lastFinishedPulling="2025-11-24 10:55:16.467991959 +0000 UTC m=+7377.002432421" observedRunningTime="2025-11-24 10:55:17.135168859 +0000 UTC m=+7377.669609321" watchObservedRunningTime="2025-11-24 10:55:17.137926627 +0000 UTC m=+7377.672367089" Nov 24 10:56:19 crc kubenswrapper[4944]: I1124 10:56:19.722324 4944 generic.go:334] "Generic (PLEG): container finished" podID="1362e83f-d5d5-43c2-bba1-21fa8ec0833e" containerID="74c022fe962256cd61e8018ba3318d965a79ac9eb2e6d6d7a20a43a4b005ec50" exitCode=0 Nov 24 10:56:19 crc kubenswrapper[4944]: I1124 10:56:19.722389 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-xmptp" event={"ID":"1362e83f-d5d5-43c2-bba1-21fa8ec0833e","Type":"ContainerDied","Data":"74c022fe962256cd61e8018ba3318d965a79ac9eb2e6d6d7a20a43a4b005ec50"} Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.213976 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.390174 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ceph\") pod \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.390597 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-inventory\") pod \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.390645 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t8zt\" (UniqueName: \"kubernetes.io/projected/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-kube-api-access-8t8zt\") pod \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.390670 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ssh-key\") pod \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.390709 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovncontroller-config-0\") pod \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.390814 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovn-combined-ca-bundle\") pod \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\" (UID: \"1362e83f-d5d5-43c2-bba1-21fa8ec0833e\") " Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.401638 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-kube-api-access-8t8zt" (OuterVolumeSpecName: "kube-api-access-8t8zt") pod "1362e83f-d5d5-43c2-bba1-21fa8ec0833e" (UID: "1362e83f-d5d5-43c2-bba1-21fa8ec0833e"). InnerVolumeSpecName "kube-api-access-8t8zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.404386 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ceph" (OuterVolumeSpecName: "ceph") pod "1362e83f-d5d5-43c2-bba1-21fa8ec0833e" (UID: "1362e83f-d5d5-43c2-bba1-21fa8ec0833e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.404449 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "1362e83f-d5d5-43c2-bba1-21fa8ec0833e" (UID: "1362e83f-d5d5-43c2-bba1-21fa8ec0833e"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.420946 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1362e83f-d5d5-43c2-bba1-21fa8ec0833e" (UID: "1362e83f-d5d5-43c2-bba1-21fa8ec0833e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.429015 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "1362e83f-d5d5-43c2-bba1-21fa8ec0833e" (UID: "1362e83f-d5d5-43c2-bba1-21fa8ec0833e"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.438668 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-inventory" (OuterVolumeSpecName: "inventory") pod "1362e83f-d5d5-43c2-bba1-21fa8ec0833e" (UID: "1362e83f-d5d5-43c2-bba1-21fa8ec0833e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.493912 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.493947 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.493958 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t8zt\" (UniqueName: \"kubernetes.io/projected/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-kube-api-access-8t8zt\") on node \"crc\" DevicePath \"\"" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.493968 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.493978 4944 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.493986 4944 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1362e83f-d5d5-43c2-bba1-21fa8ec0833e-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.744408 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-xmptp" event={"ID":"1362e83f-d5d5-43c2-bba1-21fa8ec0833e","Type":"ContainerDied","Data":"e15916b3fd2e63ca6530ca214da5d8df5dd6429333a8253a7885f30ef038c535"} Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.744444 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-xmptp" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.744446 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e15916b3fd2e63ca6530ca214da5d8df5dd6429333a8253a7885f30ef038c535" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.846817 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-sp7hl"] Nov 24 10:56:21 crc kubenswrapper[4944]: E1124 10:56:21.847796 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1362e83f-d5d5-43c2-bba1-21fa8ec0833e" containerName="ovn-openstack-openstack-cell1" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.847838 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1362e83f-d5d5-43c2-bba1-21fa8ec0833e" containerName="ovn-openstack-openstack-cell1" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.848461 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1362e83f-d5d5-43c2-bba1-21fa8ec0833e" containerName="ovn-openstack-openstack-cell1" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.850158 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.853532 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.854309 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.854774 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.854827 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.854791 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.855117 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:56:21 crc kubenswrapper[4944]: I1124 10:56:21.855663 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-sp7hl"] Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.001972 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrd8k\" (UniqueName: \"kubernetes.io/projected/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-kube-api-access-mrd8k\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.002353 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.002393 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.002462 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.002494 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.002550 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.002701 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.105109 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.105209 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.105316 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrd8k\" (UniqueName: \"kubernetes.io/projected/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-kube-api-access-mrd8k\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.105343 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.105365 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.105412 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.105429 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.110400 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.110715 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.110802 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.114330 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.116739 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.117395 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.123241 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrd8k\" (UniqueName: \"kubernetes.io/projected/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-kube-api-access-mrd8k\") pod \"neutron-metadata-openstack-openstack-cell1-sp7hl\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.176504 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:56:22 crc kubenswrapper[4944]: I1124 10:56:22.787706 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-sp7hl"] Nov 24 10:56:23 crc kubenswrapper[4944]: I1124 10:56:23.763411 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" event={"ID":"ed64daec-13f6-4d5f-8f9b-f1cd983d5727","Type":"ContainerStarted","Data":"418d1bb9c210f1a173a6b639c810a348e414e7ae6b4d024a3cb8e464e4c42f86"} Nov 24 10:56:24 crc kubenswrapper[4944]: I1124 10:56:24.773986 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" event={"ID":"ed64daec-13f6-4d5f-8f9b-f1cd983d5727","Type":"ContainerStarted","Data":"aa643aa663d414ace209b490d74fd45063ee30a0510c5e325471b77eddd7428b"} Nov 24 10:56:24 crc kubenswrapper[4944]: I1124 10:56:24.792973 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" podStartSLOduration=2.884131408 podStartE2EDuration="3.792957017s" podCreationTimestamp="2025-11-24 10:56:21 +0000 UTC" firstStartedPulling="2025-11-24 10:56:22.795970461 +0000 UTC m=+7443.330410923" lastFinishedPulling="2025-11-24 10:56:23.70479607 +0000 UTC m=+7444.239236532" observedRunningTime="2025-11-24 10:56:24.78927901 +0000 UTC m=+7445.323719482" watchObservedRunningTime="2025-11-24 10:56:24.792957017 +0000 UTC m=+7445.327397479" Nov 24 10:57:16 crc kubenswrapper[4944]: I1124 10:57:16.908214 4944 generic.go:334] "Generic (PLEG): container finished" podID="ed64daec-13f6-4d5f-8f9b-f1cd983d5727" containerID="aa643aa663d414ace209b490d74fd45063ee30a0510c5e325471b77eddd7428b" exitCode=0 Nov 24 10:57:16 crc kubenswrapper[4944]: I1124 10:57:16.908612 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" event={"ID":"ed64daec-13f6-4d5f-8f9b-f1cd983d5727","Type":"ContainerDied","Data":"aa643aa663d414ace209b490d74fd45063ee30a0510c5e325471b77eddd7428b"} Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.374097 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.505739 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-metadata-combined-ca-bundle\") pod \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.505820 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrd8k\" (UniqueName: \"kubernetes.io/projected/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-kube-api-access-mrd8k\") pod \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.505997 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-nova-metadata-neutron-config-0\") pod \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.506066 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ssh-key\") pod \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.506148 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ceph\") pod \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.506174 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-inventory\") pod \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.506251 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-ovn-metadata-agent-neutron-config-0\") pod \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\" (UID: \"ed64daec-13f6-4d5f-8f9b-f1cd983d5727\") " Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.511469 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ceph" (OuterVolumeSpecName: "ceph") pod "ed64daec-13f6-4d5f-8f9b-f1cd983d5727" (UID: "ed64daec-13f6-4d5f-8f9b-f1cd983d5727"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.517946 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "ed64daec-13f6-4d5f-8f9b-f1cd983d5727" (UID: "ed64daec-13f6-4d5f-8f9b-f1cd983d5727"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.518009 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-kube-api-access-mrd8k" (OuterVolumeSpecName: "kube-api-access-mrd8k") pod "ed64daec-13f6-4d5f-8f9b-f1cd983d5727" (UID: "ed64daec-13f6-4d5f-8f9b-f1cd983d5727"). InnerVolumeSpecName "kube-api-access-mrd8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.534998 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-inventory" (OuterVolumeSpecName: "inventory") pod "ed64daec-13f6-4d5f-8f9b-f1cd983d5727" (UID: "ed64daec-13f6-4d5f-8f9b-f1cd983d5727"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.535191 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "ed64daec-13f6-4d5f-8f9b-f1cd983d5727" (UID: "ed64daec-13f6-4d5f-8f9b-f1cd983d5727"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.538787 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "ed64daec-13f6-4d5f-8f9b-f1cd983d5727" (UID: "ed64daec-13f6-4d5f-8f9b-f1cd983d5727"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.539247 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ed64daec-13f6-4d5f-8f9b-f1cd983d5727" (UID: "ed64daec-13f6-4d5f-8f9b-f1cd983d5727"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.609344 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.609377 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.609389 4944 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.609399 4944 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.609411 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrd8k\" (UniqueName: \"kubernetes.io/projected/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-kube-api-access-mrd8k\") on node \"crc\" DevicePath \"\"" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.609420 4944 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.609429 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed64daec-13f6-4d5f-8f9b-f1cd983d5727-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.934012 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" event={"ID":"ed64daec-13f6-4d5f-8f9b-f1cd983d5727","Type":"ContainerDied","Data":"418d1bb9c210f1a173a6b639c810a348e414e7ae6b4d024a3cb8e464e4c42f86"} Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.934314 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="418d1bb9c210f1a173a6b639c810a348e414e7ae6b4d024a3cb8e464e4c42f86" Nov 24 10:57:18 crc kubenswrapper[4944]: I1124 10:57:18.934131 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-sp7hl" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.012658 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-2nlwg"] Nov 24 10:57:19 crc kubenswrapper[4944]: E1124 10:57:19.013131 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed64daec-13f6-4d5f-8f9b-f1cd983d5727" containerName="neutron-metadata-openstack-openstack-cell1" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.013152 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed64daec-13f6-4d5f-8f9b-f1cd983d5727" containerName="neutron-metadata-openstack-openstack-cell1" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.013332 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed64daec-13f6-4d5f-8f9b-f1cd983d5727" containerName="neutron-metadata-openstack-openstack-cell1" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.014036 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.017188 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.017202 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.017324 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.017536 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.017633 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.028973 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-2nlwg"] Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.120646 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.120825 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-inventory\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.120854 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ceph\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.120886 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.122167 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv7hk\" (UniqueName: \"kubernetes.io/projected/2436e674-f08d-40d2-9c86-79e807205863-kube-api-access-lv7hk\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.123286 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ssh-key\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.226067 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-inventory\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.226118 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ceph\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.226155 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.226235 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv7hk\" (UniqueName: \"kubernetes.io/projected/2436e674-f08d-40d2-9c86-79e807205863-kube-api-access-lv7hk\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.226411 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ssh-key\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.226458 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.231382 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ceph\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.232387 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.232602 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.232659 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ssh-key\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.234343 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-inventory\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.253950 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv7hk\" (UniqueName: \"kubernetes.io/projected/2436e674-f08d-40d2-9c86-79e807205863-kube-api-access-lv7hk\") pod \"libvirt-openstack-openstack-cell1-2nlwg\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.366858 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.888983 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-2nlwg"] Nov 24 10:57:19 crc kubenswrapper[4944]: I1124 10:57:19.946969 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" event={"ID":"2436e674-f08d-40d2-9c86-79e807205863","Type":"ContainerStarted","Data":"23815a4a5da5bdd83b2eb57a2547d0ca83a7555c69287b7f38e2dbc5719109c3"} Nov 24 10:57:20 crc kubenswrapper[4944]: I1124 10:57:20.606271 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 10:57:20 crc kubenswrapper[4944]: I1124 10:57:20.960351 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" event={"ID":"2436e674-f08d-40d2-9c86-79e807205863","Type":"ContainerStarted","Data":"cd916f3f017bf2fea25de28f580180e7ff9afb5838de33e3918dc4ac5cf8d31a"} Nov 24 10:57:20 crc kubenswrapper[4944]: I1124 10:57:20.983379 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" podStartSLOduration=2.273330826 podStartE2EDuration="2.983360757s" podCreationTimestamp="2025-11-24 10:57:18 +0000 UTC" firstStartedPulling="2025-11-24 10:57:19.893543297 +0000 UTC m=+7500.427983759" lastFinishedPulling="2025-11-24 10:57:20.603573208 +0000 UTC m=+7501.138013690" observedRunningTime="2025-11-24 10:57:20.976572771 +0000 UTC m=+7501.511013243" watchObservedRunningTime="2025-11-24 10:57:20.983360757 +0000 UTC m=+7501.517801209" Nov 24 10:57:23 crc kubenswrapper[4944]: I1124 10:57:23.548478 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:57:23 crc kubenswrapper[4944]: I1124 10:57:23.549128 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:57:32 crc kubenswrapper[4944]: I1124 10:57:32.777667 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5l6nk"] Nov 24 10:57:32 crc kubenswrapper[4944]: I1124 10:57:32.780297 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:32 crc kubenswrapper[4944]: I1124 10:57:32.813236 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5l6nk"] Nov 24 10:57:32 crc kubenswrapper[4944]: I1124 10:57:32.971744 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-catalog-content\") pod \"redhat-marketplace-5l6nk\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:32 crc kubenswrapper[4944]: I1124 10:57:32.971795 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84lt5\" (UniqueName: \"kubernetes.io/projected/82889d4e-3ff0-449f-a06a-e4131cf473a0-kube-api-access-84lt5\") pod \"redhat-marketplace-5l6nk\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:32 crc kubenswrapper[4944]: I1124 10:57:32.971963 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-utilities\") pod \"redhat-marketplace-5l6nk\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:33 crc kubenswrapper[4944]: I1124 10:57:33.074313 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-utilities\") pod \"redhat-marketplace-5l6nk\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:33 crc kubenswrapper[4944]: I1124 10:57:33.074434 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-catalog-content\") pod \"redhat-marketplace-5l6nk\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:33 crc kubenswrapper[4944]: I1124 10:57:33.074463 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84lt5\" (UniqueName: \"kubernetes.io/projected/82889d4e-3ff0-449f-a06a-e4131cf473a0-kube-api-access-84lt5\") pod \"redhat-marketplace-5l6nk\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:33 crc kubenswrapper[4944]: I1124 10:57:33.074845 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-utilities\") pod \"redhat-marketplace-5l6nk\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:33 crc kubenswrapper[4944]: I1124 10:57:33.074915 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-catalog-content\") pod \"redhat-marketplace-5l6nk\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:33 crc kubenswrapper[4944]: I1124 10:57:33.099173 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84lt5\" (UniqueName: \"kubernetes.io/projected/82889d4e-3ff0-449f-a06a-e4131cf473a0-kube-api-access-84lt5\") pod \"redhat-marketplace-5l6nk\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:33 crc kubenswrapper[4944]: I1124 10:57:33.155820 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:33 crc kubenswrapper[4944]: I1124 10:57:33.669301 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5l6nk"] Nov 24 10:57:34 crc kubenswrapper[4944]: I1124 10:57:34.099535 4944 generic.go:334] "Generic (PLEG): container finished" podID="82889d4e-3ff0-449f-a06a-e4131cf473a0" containerID="5617fa3a1bb6e00dd009ab65036aa41b0ac773c6ae318a3d5485bd8db3a1c28b" exitCode=0 Nov 24 10:57:34 crc kubenswrapper[4944]: I1124 10:57:34.099588 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5l6nk" event={"ID":"82889d4e-3ff0-449f-a06a-e4131cf473a0","Type":"ContainerDied","Data":"5617fa3a1bb6e00dd009ab65036aa41b0ac773c6ae318a3d5485bd8db3a1c28b"} Nov 24 10:57:34 crc kubenswrapper[4944]: I1124 10:57:34.099637 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5l6nk" event={"ID":"82889d4e-3ff0-449f-a06a-e4131cf473a0","Type":"ContainerStarted","Data":"1d922138bd1511b5d592087a61fa9f6c868d1609223b1f9218da5d8da70a5e68"} Nov 24 10:57:35 crc kubenswrapper[4944]: I1124 10:57:35.110878 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5l6nk" event={"ID":"82889d4e-3ff0-449f-a06a-e4131cf473a0","Type":"ContainerStarted","Data":"e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2"} Nov 24 10:57:36 crc kubenswrapper[4944]: I1124 10:57:36.129946 4944 generic.go:334] "Generic (PLEG): container finished" podID="82889d4e-3ff0-449f-a06a-e4131cf473a0" containerID="e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2" exitCode=0 Nov 24 10:57:36 crc kubenswrapper[4944]: I1124 10:57:36.130325 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5l6nk" event={"ID":"82889d4e-3ff0-449f-a06a-e4131cf473a0","Type":"ContainerDied","Data":"e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2"} Nov 24 10:57:37 crc kubenswrapper[4944]: I1124 10:57:37.142072 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5l6nk" event={"ID":"82889d4e-3ff0-449f-a06a-e4131cf473a0","Type":"ContainerStarted","Data":"53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85"} Nov 24 10:57:37 crc kubenswrapper[4944]: I1124 10:57:37.163827 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5l6nk" podStartSLOduration=2.724355297 podStartE2EDuration="5.163808742s" podCreationTimestamp="2025-11-24 10:57:32 +0000 UTC" firstStartedPulling="2025-11-24 10:57:34.102955392 +0000 UTC m=+7514.637395864" lastFinishedPulling="2025-11-24 10:57:36.542408847 +0000 UTC m=+7517.076849309" observedRunningTime="2025-11-24 10:57:37.156682106 +0000 UTC m=+7517.691122578" watchObservedRunningTime="2025-11-24 10:57:37.163808742 +0000 UTC m=+7517.698249204" Nov 24 10:57:43 crc kubenswrapper[4944]: I1124 10:57:43.156381 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:43 crc kubenswrapper[4944]: I1124 10:57:43.156919 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:43 crc kubenswrapper[4944]: I1124 10:57:43.220902 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:43 crc kubenswrapper[4944]: I1124 10:57:43.280668 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:43 crc kubenswrapper[4944]: I1124 10:57:43.459169 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5l6nk"] Nov 24 10:57:45 crc kubenswrapper[4944]: I1124 10:57:45.222027 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5l6nk" podUID="82889d4e-3ff0-449f-a06a-e4131cf473a0" containerName="registry-server" containerID="cri-o://53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85" gracePeriod=2 Nov 24 10:57:45 crc kubenswrapper[4944]: I1124 10:57:45.750851 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:45 crc kubenswrapper[4944]: I1124 10:57:45.761864 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84lt5\" (UniqueName: \"kubernetes.io/projected/82889d4e-3ff0-449f-a06a-e4131cf473a0-kube-api-access-84lt5\") pod \"82889d4e-3ff0-449f-a06a-e4131cf473a0\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " Nov 24 10:57:45 crc kubenswrapper[4944]: I1124 10:57:45.762084 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-catalog-content\") pod \"82889d4e-3ff0-449f-a06a-e4131cf473a0\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " Nov 24 10:57:45 crc kubenswrapper[4944]: I1124 10:57:45.762238 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-utilities\") pod \"82889d4e-3ff0-449f-a06a-e4131cf473a0\" (UID: \"82889d4e-3ff0-449f-a06a-e4131cf473a0\") " Nov 24 10:57:45 crc kubenswrapper[4944]: I1124 10:57:45.764025 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-utilities" (OuterVolumeSpecName: "utilities") pod "82889d4e-3ff0-449f-a06a-e4131cf473a0" (UID: "82889d4e-3ff0-449f-a06a-e4131cf473a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:57:45 crc kubenswrapper[4944]: I1124 10:57:45.768191 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82889d4e-3ff0-449f-a06a-e4131cf473a0-kube-api-access-84lt5" (OuterVolumeSpecName: "kube-api-access-84lt5") pod "82889d4e-3ff0-449f-a06a-e4131cf473a0" (UID: "82889d4e-3ff0-449f-a06a-e4131cf473a0"). InnerVolumeSpecName "kube-api-access-84lt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 10:57:45 crc kubenswrapper[4944]: I1124 10:57:45.785463 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82889d4e-3ff0-449f-a06a-e4131cf473a0" (UID: "82889d4e-3ff0-449f-a06a-e4131cf473a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 10:57:45 crc kubenswrapper[4944]: I1124 10:57:45.864775 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 10:57:45 crc kubenswrapper[4944]: I1124 10:57:45.864818 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84lt5\" (UniqueName: \"kubernetes.io/projected/82889d4e-3ff0-449f-a06a-e4131cf473a0-kube-api-access-84lt5\") on node \"crc\" DevicePath \"\"" Nov 24 10:57:45 crc kubenswrapper[4944]: I1124 10:57:45.864833 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82889d4e-3ff0-449f-a06a-e4131cf473a0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.233658 4944 generic.go:334] "Generic (PLEG): container finished" podID="82889d4e-3ff0-449f-a06a-e4131cf473a0" containerID="53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85" exitCode=0 Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.233712 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5l6nk" event={"ID":"82889d4e-3ff0-449f-a06a-e4131cf473a0","Type":"ContainerDied","Data":"53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85"} Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.233757 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5l6nk" Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.233776 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5l6nk" event={"ID":"82889d4e-3ff0-449f-a06a-e4131cf473a0","Type":"ContainerDied","Data":"1d922138bd1511b5d592087a61fa9f6c868d1609223b1f9218da5d8da70a5e68"} Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.233869 4944 scope.go:117] "RemoveContainer" containerID="53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85" Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.259294 4944 scope.go:117] "RemoveContainer" containerID="e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2" Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.267356 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5l6nk"] Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.286078 4944 scope.go:117] "RemoveContainer" containerID="5617fa3a1bb6e00dd009ab65036aa41b0ac773c6ae318a3d5485bd8db3a1c28b" Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.292661 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5l6nk"] Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.309355 4944 scope.go:117] "RemoveContainer" containerID="53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85" Nov 24 10:57:46 crc kubenswrapper[4944]: E1124 10:57:46.312670 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85\": container with ID starting with 53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85 not found: ID does not exist" containerID="53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85" Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.312744 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85"} err="failed to get container status \"53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85\": rpc error: code = NotFound desc = could not find container \"53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85\": container with ID starting with 53decfd3984b874a6507d85c91119451328b7b63932c058b1a5a18de8b047a85 not found: ID does not exist" Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.312781 4944 scope.go:117] "RemoveContainer" containerID="e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2" Nov 24 10:57:46 crc kubenswrapper[4944]: E1124 10:57:46.313282 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2\": container with ID starting with e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2 not found: ID does not exist" containerID="e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2" Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.313331 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2"} err="failed to get container status \"e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2\": rpc error: code = NotFound desc = could not find container \"e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2\": container with ID starting with e5224782b65534bdb5f9a3c934b2fa7dabfbf47dc020560a4776985f886f6ea2 not found: ID does not exist" Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.313349 4944 scope.go:117] "RemoveContainer" containerID="5617fa3a1bb6e00dd009ab65036aa41b0ac773c6ae318a3d5485bd8db3a1c28b" Nov 24 10:57:46 crc kubenswrapper[4944]: E1124 10:57:46.313675 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5617fa3a1bb6e00dd009ab65036aa41b0ac773c6ae318a3d5485bd8db3a1c28b\": container with ID starting with 5617fa3a1bb6e00dd009ab65036aa41b0ac773c6ae318a3d5485bd8db3a1c28b not found: ID does not exist" containerID="5617fa3a1bb6e00dd009ab65036aa41b0ac773c6ae318a3d5485bd8db3a1c28b" Nov 24 10:57:46 crc kubenswrapper[4944]: I1124 10:57:46.313693 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5617fa3a1bb6e00dd009ab65036aa41b0ac773c6ae318a3d5485bd8db3a1c28b"} err="failed to get container status \"5617fa3a1bb6e00dd009ab65036aa41b0ac773c6ae318a3d5485bd8db3a1c28b\": rpc error: code = NotFound desc = could not find container \"5617fa3a1bb6e00dd009ab65036aa41b0ac773c6ae318a3d5485bd8db3a1c28b\": container with ID starting with 5617fa3a1bb6e00dd009ab65036aa41b0ac773c6ae318a3d5485bd8db3a1c28b not found: ID does not exist" Nov 24 10:57:48 crc kubenswrapper[4944]: I1124 10:57:48.320415 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82889d4e-3ff0-449f-a06a-e4131cf473a0" path="/var/lib/kubelet/pods/82889d4e-3ff0-449f-a06a-e4131cf473a0/volumes" Nov 24 10:57:53 crc kubenswrapper[4944]: I1124 10:57:53.549068 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:57:53 crc kubenswrapper[4944]: I1124 10:57:53.549628 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:58:23 crc kubenswrapper[4944]: I1124 10:58:23.549554 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 10:58:23 crc kubenswrapper[4944]: I1124 10:58:23.550793 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 10:58:23 crc kubenswrapper[4944]: I1124 10:58:23.550875 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 10:58:23 crc kubenswrapper[4944]: I1124 10:58:23.552543 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"711f3ecca590ed13334ef816a5b7121fb9369df33f5c8c0821dcbd83e7e9d442"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 10:58:23 crc kubenswrapper[4944]: I1124 10:58:23.552668 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://711f3ecca590ed13334ef816a5b7121fb9369df33f5c8c0821dcbd83e7e9d442" gracePeriod=600 Nov 24 10:58:24 crc kubenswrapper[4944]: I1124 10:58:24.609182 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="711f3ecca590ed13334ef816a5b7121fb9369df33f5c8c0821dcbd83e7e9d442" exitCode=0 Nov 24 10:58:24 crc kubenswrapper[4944]: I1124 10:58:24.609261 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"711f3ecca590ed13334ef816a5b7121fb9369df33f5c8c0821dcbd83e7e9d442"} Nov 24 10:58:24 crc kubenswrapper[4944]: I1124 10:58:24.609779 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2"} Nov 24 10:58:24 crc kubenswrapper[4944]: I1124 10:58:24.609815 4944 scope.go:117] "RemoveContainer" containerID="a9c68fa5d7176df60b921240911314c7ccdc9a8a38483971184fd61e27c43053" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.174744 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n76ff"] Nov 24 10:59:58 crc kubenswrapper[4944]: E1124 10:59:58.176061 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82889d4e-3ff0-449f-a06a-e4131cf473a0" containerName="registry-server" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.176079 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="82889d4e-3ff0-449f-a06a-e4131cf473a0" containerName="registry-server" Nov 24 10:59:58 crc kubenswrapper[4944]: E1124 10:59:58.176106 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82889d4e-3ff0-449f-a06a-e4131cf473a0" containerName="extract-utilities" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.176114 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="82889d4e-3ff0-449f-a06a-e4131cf473a0" containerName="extract-utilities" Nov 24 10:59:58 crc kubenswrapper[4944]: E1124 10:59:58.176131 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82889d4e-3ff0-449f-a06a-e4131cf473a0" containerName="extract-content" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.176139 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="82889d4e-3ff0-449f-a06a-e4131cf473a0" containerName="extract-content" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.176448 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="82889d4e-3ff0-449f-a06a-e4131cf473a0" containerName="registry-server" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.178675 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n76ff" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.187530 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n76ff"] Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.358598 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swwdj\" (UniqueName: \"kubernetes.io/projected/4abeaa40-20bf-48d8-a15e-5b81ded7b182-kube-api-access-swwdj\") pod \"community-operators-n76ff\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " pod="openshift-marketplace/community-operators-n76ff" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.358781 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-utilities\") pod \"community-operators-n76ff\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " pod="openshift-marketplace/community-operators-n76ff" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.358863 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-catalog-content\") pod \"community-operators-n76ff\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " pod="openshift-marketplace/community-operators-n76ff" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.461093 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-catalog-content\") pod \"community-operators-n76ff\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " pod="openshift-marketplace/community-operators-n76ff" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.461423 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swwdj\" (UniqueName: \"kubernetes.io/projected/4abeaa40-20bf-48d8-a15e-5b81ded7b182-kube-api-access-swwdj\") pod \"community-operators-n76ff\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " pod="openshift-marketplace/community-operators-n76ff" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.461651 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-utilities\") pod \"community-operators-n76ff\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " pod="openshift-marketplace/community-operators-n76ff" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.461682 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-catalog-content\") pod \"community-operators-n76ff\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " pod="openshift-marketplace/community-operators-n76ff" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.462711 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-utilities\") pod \"community-operators-n76ff\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " pod="openshift-marketplace/community-operators-n76ff" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.491337 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swwdj\" (UniqueName: \"kubernetes.io/projected/4abeaa40-20bf-48d8-a15e-5b81ded7b182-kube-api-access-swwdj\") pod \"community-operators-n76ff\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " pod="openshift-marketplace/community-operators-n76ff" Nov 24 10:59:58 crc kubenswrapper[4944]: I1124 10:59:58.527083 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n76ff" Nov 24 10:59:59 crc kubenswrapper[4944]: I1124 10:59:59.073610 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n76ff"] Nov 24 10:59:59 crc kubenswrapper[4944]: I1124 10:59:59.603257 4944 generic.go:334] "Generic (PLEG): container finished" podID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" containerID="368cdf75206c09dd1fa3de9ea9d8c02c958067ea9a8d192c6847bf926f71225b" exitCode=0 Nov 24 10:59:59 crc kubenswrapper[4944]: I1124 10:59:59.603379 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n76ff" event={"ID":"4abeaa40-20bf-48d8-a15e-5b81ded7b182","Type":"ContainerDied","Data":"368cdf75206c09dd1fa3de9ea9d8c02c958067ea9a8d192c6847bf926f71225b"} Nov 24 10:59:59 crc kubenswrapper[4944]: I1124 10:59:59.603675 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n76ff" event={"ID":"4abeaa40-20bf-48d8-a15e-5b81ded7b182","Type":"ContainerStarted","Data":"9304bc496f7ffa3420c75c73b0dd2779f2bd0508887fa25e617f5a9199cf67fa"} Nov 24 10:59:59 crc kubenswrapper[4944]: I1124 10:59:59.605118 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.156797 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv"] Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.169705 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.172527 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv"] Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.175491 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.175863 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.308219 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-secret-volume\") pod \"collect-profiles-29399700-8q6rv\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.308653 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-config-volume\") pod \"collect-profiles-29399700-8q6rv\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.309149 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qknv\" (UniqueName: \"kubernetes.io/projected/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-kube-api-access-5qknv\") pod \"collect-profiles-29399700-8q6rv\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.411199 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qknv\" (UniqueName: \"kubernetes.io/projected/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-kube-api-access-5qknv\") pod \"collect-profiles-29399700-8q6rv\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.411291 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-secret-volume\") pod \"collect-profiles-29399700-8q6rv\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.411375 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-config-volume\") pod \"collect-profiles-29399700-8q6rv\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.412557 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-config-volume\") pod \"collect-profiles-29399700-8q6rv\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.418617 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-secret-volume\") pod \"collect-profiles-29399700-8q6rv\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.428429 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qknv\" (UniqueName: \"kubernetes.io/projected/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-kube-api-access-5qknv\") pod \"collect-profiles-29399700-8q6rv\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.503169 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:00 crc kubenswrapper[4944]: I1124 11:00:00.964459 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv"] Nov 24 11:00:01 crc kubenswrapper[4944]: I1124 11:00:01.635827 4944 generic.go:334] "Generic (PLEG): container finished" podID="bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e" containerID="433a0c4e7a6c0666f2dc688f2c6c3d0cf83ba14f7a4d7f592598f3e2740a4324" exitCode=0 Nov 24 11:00:01 crc kubenswrapper[4944]: I1124 11:00:01.635889 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" event={"ID":"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e","Type":"ContainerDied","Data":"433a0c4e7a6c0666f2dc688f2c6c3d0cf83ba14f7a4d7f592598f3e2740a4324"} Nov 24 11:00:01 crc kubenswrapper[4944]: I1124 11:00:01.636517 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" event={"ID":"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e","Type":"ContainerStarted","Data":"b7ba560ab3a8a5ebd898b064c1f406abe2465f62e8a6339daebc509fb4639e57"} Nov 24 11:00:01 crc kubenswrapper[4944]: I1124 11:00:01.638498 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n76ff" event={"ID":"4abeaa40-20bf-48d8-a15e-5b81ded7b182","Type":"ContainerStarted","Data":"3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e"} Nov 24 11:00:02 crc kubenswrapper[4944]: I1124 11:00:02.648310 4944 generic.go:334] "Generic (PLEG): container finished" podID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" containerID="3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e" exitCode=0 Nov 24 11:00:02 crc kubenswrapper[4944]: I1124 11:00:02.648373 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n76ff" event={"ID":"4abeaa40-20bf-48d8-a15e-5b81ded7b182","Type":"ContainerDied","Data":"3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e"} Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.001930 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.168594 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-config-volume\") pod \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.168708 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qknv\" (UniqueName: \"kubernetes.io/projected/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-kube-api-access-5qknv\") pod \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.168755 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-secret-volume\") pod \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\" (UID: \"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e\") " Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.169479 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-config-volume" (OuterVolumeSpecName: "config-volume") pod "bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e" (UID: "bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.174989 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e" (UID: "bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.175068 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-kube-api-access-5qknv" (OuterVolumeSpecName: "kube-api-access-5qknv") pod "bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e" (UID: "bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e"). InnerVolumeSpecName "kube-api-access-5qknv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.271074 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.271349 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qknv\" (UniqueName: \"kubernetes.io/projected/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-kube-api-access-5qknv\") on node \"crc\" DevicePath \"\"" Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.271439 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.661391 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n76ff" event={"ID":"4abeaa40-20bf-48d8-a15e-5b81ded7b182","Type":"ContainerStarted","Data":"6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e"} Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.663636 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" event={"ID":"bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e","Type":"ContainerDied","Data":"b7ba560ab3a8a5ebd898b064c1f406abe2465f62e8a6339daebc509fb4639e57"} Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.663667 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7ba560ab3a8a5ebd898b064c1f406abe2465f62e8a6339daebc509fb4639e57" Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.663678 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399700-8q6rv" Nov 24 11:00:03 crc kubenswrapper[4944]: I1124 11:00:03.679962 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n76ff" podStartSLOduration=2.137486486 podStartE2EDuration="5.67994177s" podCreationTimestamp="2025-11-24 10:59:58 +0000 UTC" firstStartedPulling="2025-11-24 10:59:59.60482635 +0000 UTC m=+7660.139266812" lastFinishedPulling="2025-11-24 11:00:03.147281634 +0000 UTC m=+7663.681722096" observedRunningTime="2025-11-24 11:00:03.676277243 +0000 UTC m=+7664.210717725" watchObservedRunningTime="2025-11-24 11:00:03.67994177 +0000 UTC m=+7664.214382232" Nov 24 11:00:04 crc kubenswrapper[4944]: I1124 11:00:04.065665 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr"] Nov 24 11:00:04 crc kubenswrapper[4944]: I1124 11:00:04.076364 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399655-xplzr"] Nov 24 11:00:04 crc kubenswrapper[4944]: I1124 11:00:04.290671 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47ead2de-7b54-48d8-a5ec-f292678c1966" path="/var/lib/kubelet/pods/47ead2de-7b54-48d8-a5ec-f292678c1966/volumes" Nov 24 11:00:08 crc kubenswrapper[4944]: I1124 11:00:08.528029 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n76ff" Nov 24 11:00:08 crc kubenswrapper[4944]: I1124 11:00:08.528828 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n76ff" Nov 24 11:00:08 crc kubenswrapper[4944]: I1124 11:00:08.585320 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n76ff" Nov 24 11:00:08 crc kubenswrapper[4944]: I1124 11:00:08.771682 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n76ff" Nov 24 11:00:08 crc kubenswrapper[4944]: I1124 11:00:08.819258 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n76ff"] Nov 24 11:00:10 crc kubenswrapper[4944]: I1124 11:00:10.741370 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n76ff" podUID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" containerName="registry-server" containerID="cri-o://6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e" gracePeriod=2 Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.251941 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n76ff" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.447807 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-catalog-content\") pod \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.447876 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-utilities\") pod \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.448220 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swwdj\" (UniqueName: \"kubernetes.io/projected/4abeaa40-20bf-48d8-a15e-5b81ded7b182-kube-api-access-swwdj\") pod \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\" (UID: \"4abeaa40-20bf-48d8-a15e-5b81ded7b182\") " Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.449092 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-utilities" (OuterVolumeSpecName: "utilities") pod "4abeaa40-20bf-48d8-a15e-5b81ded7b182" (UID: "4abeaa40-20bf-48d8-a15e-5b81ded7b182"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.454513 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4abeaa40-20bf-48d8-a15e-5b81ded7b182-kube-api-access-swwdj" (OuterVolumeSpecName: "kube-api-access-swwdj") pod "4abeaa40-20bf-48d8-a15e-5b81ded7b182" (UID: "4abeaa40-20bf-48d8-a15e-5b81ded7b182"). InnerVolumeSpecName "kube-api-access-swwdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.497855 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4abeaa40-20bf-48d8-a15e-5b81ded7b182" (UID: "4abeaa40-20bf-48d8-a15e-5b81ded7b182"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.550334 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swwdj\" (UniqueName: \"kubernetes.io/projected/4abeaa40-20bf-48d8-a15e-5b81ded7b182-kube-api-access-swwdj\") on node \"crc\" DevicePath \"\"" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.550364 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.550375 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4abeaa40-20bf-48d8-a15e-5b81ded7b182-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.753838 4944 generic.go:334] "Generic (PLEG): container finished" podID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" containerID="6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e" exitCode=0 Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.753907 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n76ff" event={"ID":"4abeaa40-20bf-48d8-a15e-5b81ded7b182","Type":"ContainerDied","Data":"6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e"} Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.753945 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n76ff" event={"ID":"4abeaa40-20bf-48d8-a15e-5b81ded7b182","Type":"ContainerDied","Data":"9304bc496f7ffa3420c75c73b0dd2779f2bd0508887fa25e617f5a9199cf67fa"} Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.753972 4944 scope.go:117] "RemoveContainer" containerID="6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.754189 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n76ff" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.787321 4944 scope.go:117] "RemoveContainer" containerID="3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.788178 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n76ff"] Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.797267 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n76ff"] Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.812746 4944 scope.go:117] "RemoveContainer" containerID="368cdf75206c09dd1fa3de9ea9d8c02c958067ea9a8d192c6847bf926f71225b" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.854033 4944 scope.go:117] "RemoveContainer" containerID="6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e" Nov 24 11:00:11 crc kubenswrapper[4944]: E1124 11:00:11.854451 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e\": container with ID starting with 6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e not found: ID does not exist" containerID="6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.854564 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e"} err="failed to get container status \"6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e\": rpc error: code = NotFound desc = could not find container \"6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e\": container with ID starting with 6d110c1ff8e997946d3490f209f49fa8ae06e3f04eae3f8de527d58c333e684e not found: ID does not exist" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.854724 4944 scope.go:117] "RemoveContainer" containerID="3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e" Nov 24 11:00:11 crc kubenswrapper[4944]: E1124 11:00:11.855133 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e\": container with ID starting with 3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e not found: ID does not exist" containerID="3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.855210 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e"} err="failed to get container status \"3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e\": rpc error: code = NotFound desc = could not find container \"3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e\": container with ID starting with 3962cc57e8421f0cfba8356cbadcd022c665e61aaf7432406a8acca04684986e not found: ID does not exist" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.855230 4944 scope.go:117] "RemoveContainer" containerID="368cdf75206c09dd1fa3de9ea9d8c02c958067ea9a8d192c6847bf926f71225b" Nov 24 11:00:11 crc kubenswrapper[4944]: E1124 11:00:11.855495 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"368cdf75206c09dd1fa3de9ea9d8c02c958067ea9a8d192c6847bf926f71225b\": container with ID starting with 368cdf75206c09dd1fa3de9ea9d8c02c958067ea9a8d192c6847bf926f71225b not found: ID does not exist" containerID="368cdf75206c09dd1fa3de9ea9d8c02c958067ea9a8d192c6847bf926f71225b" Nov 24 11:00:11 crc kubenswrapper[4944]: I1124 11:00:11.855596 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"368cdf75206c09dd1fa3de9ea9d8c02c958067ea9a8d192c6847bf926f71225b"} err="failed to get container status \"368cdf75206c09dd1fa3de9ea9d8c02c958067ea9a8d192c6847bf926f71225b\": rpc error: code = NotFound desc = could not find container \"368cdf75206c09dd1fa3de9ea9d8c02c958067ea9a8d192c6847bf926f71225b\": container with ID starting with 368cdf75206c09dd1fa3de9ea9d8c02c958067ea9a8d192c6847bf926f71225b not found: ID does not exist" Nov 24 11:00:12 crc kubenswrapper[4944]: I1124 11:00:12.290818 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" path="/var/lib/kubelet/pods/4abeaa40-20bf-48d8-a15e-5b81ded7b182/volumes" Nov 24 11:00:36 crc kubenswrapper[4944]: I1124 11:00:36.447221 4944 scope.go:117] "RemoveContainer" containerID="a2458a1342744ce198f6787bedc601a87b08fcfa6bccf41b35d86c2aa2a442a8" Nov 24 11:00:53 crc kubenswrapper[4944]: I1124 11:00:53.549086 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:00:53 crc kubenswrapper[4944]: I1124 11:00:53.549825 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.176092 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29399701-jl9qm"] Nov 24 11:01:00 crc kubenswrapper[4944]: E1124 11:01:00.177292 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" containerName="registry-server" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.177307 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" containerName="registry-server" Nov 24 11:01:00 crc kubenswrapper[4944]: E1124 11:01:00.177315 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" containerName="extract-content" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.177321 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" containerName="extract-content" Nov 24 11:01:00 crc kubenswrapper[4944]: E1124 11:01:00.177357 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e" containerName="collect-profiles" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.177364 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e" containerName="collect-profiles" Nov 24 11:01:00 crc kubenswrapper[4944]: E1124 11:01:00.177372 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" containerName="extract-utilities" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.177378 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" containerName="extract-utilities" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.177743 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="4abeaa40-20bf-48d8-a15e-5b81ded7b182" containerName="registry-server" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.177774 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bebb4cd3-c6fe-48e8-ba8a-18dedc9adb1e" containerName="collect-profiles" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.178611 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.218667 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399701-jl9qm"] Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.259166 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-config-data\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.259298 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-combined-ca-bundle\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.259340 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pbd4\" (UniqueName: \"kubernetes.io/projected/72f27aa3-7622-4e53-bef1-89f30315ceb3-kube-api-access-5pbd4\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.259578 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-fernet-keys\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.361759 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-combined-ca-bundle\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.361841 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pbd4\" (UniqueName: \"kubernetes.io/projected/72f27aa3-7622-4e53-bef1-89f30315ceb3-kube-api-access-5pbd4\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.361952 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-fernet-keys\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.361993 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-config-data\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.368846 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-config-data\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.369107 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-fernet-keys\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.373144 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-combined-ca-bundle\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.382003 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pbd4\" (UniqueName: \"kubernetes.io/projected/72f27aa3-7622-4e53-bef1-89f30315ceb3-kube-api-access-5pbd4\") pod \"keystone-cron-29399701-jl9qm\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.538693 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:00 crc kubenswrapper[4944]: I1124 11:01:00.975907 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399701-jl9qm"] Nov 24 11:01:00 crc kubenswrapper[4944]: W1124 11:01:00.980904 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72f27aa3_7622_4e53_bef1_89f30315ceb3.slice/crio-9253c6ba693eb33c2aeab033fbd0586e8a48cd6250fc95fd2c2d626a79592bbb WatchSource:0}: Error finding container 9253c6ba693eb33c2aeab033fbd0586e8a48cd6250fc95fd2c2d626a79592bbb: Status 404 returned error can't find the container with id 9253c6ba693eb33c2aeab033fbd0586e8a48cd6250fc95fd2c2d626a79592bbb Nov 24 11:01:01 crc kubenswrapper[4944]: I1124 11:01:01.255032 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399701-jl9qm" event={"ID":"72f27aa3-7622-4e53-bef1-89f30315ceb3","Type":"ContainerStarted","Data":"027ad923c9b36be27106f583c144336af3585ced49e08b3c7b02139d21d491e7"} Nov 24 11:01:01 crc kubenswrapper[4944]: I1124 11:01:01.255412 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399701-jl9qm" event={"ID":"72f27aa3-7622-4e53-bef1-89f30315ceb3","Type":"ContainerStarted","Data":"9253c6ba693eb33c2aeab033fbd0586e8a48cd6250fc95fd2c2d626a79592bbb"} Nov 24 11:01:01 crc kubenswrapper[4944]: I1124 11:01:01.273533 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29399701-jl9qm" podStartSLOduration=1.273516187 podStartE2EDuration="1.273516187s" podCreationTimestamp="2025-11-24 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 11:01:01.268873109 +0000 UTC m=+7721.803313581" watchObservedRunningTime="2025-11-24 11:01:01.273516187 +0000 UTC m=+7721.807956649" Nov 24 11:01:04 crc kubenswrapper[4944]: I1124 11:01:04.287124 4944 generic.go:334] "Generic (PLEG): container finished" podID="72f27aa3-7622-4e53-bef1-89f30315ceb3" containerID="027ad923c9b36be27106f583c144336af3585ced49e08b3c7b02139d21d491e7" exitCode=0 Nov 24 11:01:04 crc kubenswrapper[4944]: I1124 11:01:04.298300 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399701-jl9qm" event={"ID":"72f27aa3-7622-4e53-bef1-89f30315ceb3","Type":"ContainerDied","Data":"027ad923c9b36be27106f583c144336af3585ced49e08b3c7b02139d21d491e7"} Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.687559 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.875492 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-config-data\") pod \"72f27aa3-7622-4e53-bef1-89f30315ceb3\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.875671 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-fernet-keys\") pod \"72f27aa3-7622-4e53-bef1-89f30315ceb3\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.875705 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-combined-ca-bundle\") pod \"72f27aa3-7622-4e53-bef1-89f30315ceb3\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.875815 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pbd4\" (UniqueName: \"kubernetes.io/projected/72f27aa3-7622-4e53-bef1-89f30315ceb3-kube-api-access-5pbd4\") pod \"72f27aa3-7622-4e53-bef1-89f30315ceb3\" (UID: \"72f27aa3-7622-4e53-bef1-89f30315ceb3\") " Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.881290 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "72f27aa3-7622-4e53-bef1-89f30315ceb3" (UID: "72f27aa3-7622-4e53-bef1-89f30315ceb3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.882466 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72f27aa3-7622-4e53-bef1-89f30315ceb3-kube-api-access-5pbd4" (OuterVolumeSpecName: "kube-api-access-5pbd4") pod "72f27aa3-7622-4e53-bef1-89f30315ceb3" (UID: "72f27aa3-7622-4e53-bef1-89f30315ceb3"). InnerVolumeSpecName "kube-api-access-5pbd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.904639 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72f27aa3-7622-4e53-bef1-89f30315ceb3" (UID: "72f27aa3-7622-4e53-bef1-89f30315ceb3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.932750 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-config-data" (OuterVolumeSpecName: "config-data") pod "72f27aa3-7622-4e53-bef1-89f30315ceb3" (UID: "72f27aa3-7622-4e53-bef1-89f30315ceb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.978491 4944 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.978521 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.978531 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pbd4\" (UniqueName: \"kubernetes.io/projected/72f27aa3-7622-4e53-bef1-89f30315ceb3-kube-api-access-5pbd4\") on node \"crc\" DevicePath \"\"" Nov 24 11:01:05 crc kubenswrapper[4944]: I1124 11:01:05.978540 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72f27aa3-7622-4e53-bef1-89f30315ceb3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 11:01:06 crc kubenswrapper[4944]: I1124 11:01:06.309389 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399701-jl9qm" event={"ID":"72f27aa3-7622-4e53-bef1-89f30315ceb3","Type":"ContainerDied","Data":"9253c6ba693eb33c2aeab033fbd0586e8a48cd6250fc95fd2c2d626a79592bbb"} Nov 24 11:01:06 crc kubenswrapper[4944]: I1124 11:01:06.309421 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9253c6ba693eb33c2aeab033fbd0586e8a48cd6250fc95fd2c2d626a79592bbb" Nov 24 11:01:06 crc kubenswrapper[4944]: I1124 11:01:06.309449 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399701-jl9qm" Nov 24 11:01:23 crc kubenswrapper[4944]: I1124 11:01:23.549079 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:01:23 crc kubenswrapper[4944]: I1124 11:01:23.549719 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:01:53 crc kubenswrapper[4944]: I1124 11:01:53.548166 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:01:53 crc kubenswrapper[4944]: I1124 11:01:53.548786 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:01:53 crc kubenswrapper[4944]: I1124 11:01:53.548849 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 11:01:53 crc kubenswrapper[4944]: I1124 11:01:53.549993 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 11:01:53 crc kubenswrapper[4944]: I1124 11:01:53.550123 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" gracePeriod=600 Nov 24 11:01:53 crc kubenswrapper[4944]: E1124 11:01:53.675587 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:01:53 crc kubenswrapper[4944]: I1124 11:01:53.748459 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" exitCode=0 Nov 24 11:01:53 crc kubenswrapper[4944]: I1124 11:01:53.748499 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2"} Nov 24 11:01:53 crc kubenswrapper[4944]: I1124 11:01:53.748532 4944 scope.go:117] "RemoveContainer" containerID="711f3ecca590ed13334ef816a5b7121fb9369df33f5c8c0821dcbd83e7e9d442" Nov 24 11:01:53 crc kubenswrapper[4944]: I1124 11:01:53.749341 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:01:53 crc kubenswrapper[4944]: E1124 11:01:53.749711 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:02:05 crc kubenswrapper[4944]: I1124 11:02:05.875522 4944 generic.go:334] "Generic (PLEG): container finished" podID="2436e674-f08d-40d2-9c86-79e807205863" containerID="cd916f3f017bf2fea25de28f580180e7ff9afb5838de33e3918dc4ac5cf8d31a" exitCode=0 Nov 24 11:02:05 crc kubenswrapper[4944]: I1124 11:02:05.875648 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" event={"ID":"2436e674-f08d-40d2-9c86-79e807205863","Type":"ContainerDied","Data":"cd916f3f017bf2fea25de28f580180e7ff9afb5838de33e3918dc4ac5cf8d31a"} Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.277970 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:02:07 crc kubenswrapper[4944]: E1124 11:02:07.278512 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.408501 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.600722 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-combined-ca-bundle\") pod \"2436e674-f08d-40d2-9c86-79e807205863\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.600775 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ceph\") pod \"2436e674-f08d-40d2-9c86-79e807205863\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.600840 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-secret-0\") pod \"2436e674-f08d-40d2-9c86-79e807205863\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.600895 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv7hk\" (UniqueName: \"kubernetes.io/projected/2436e674-f08d-40d2-9c86-79e807205863-kube-api-access-lv7hk\") pod \"2436e674-f08d-40d2-9c86-79e807205863\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.600918 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ssh-key\") pod \"2436e674-f08d-40d2-9c86-79e807205863\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.601058 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-inventory\") pod \"2436e674-f08d-40d2-9c86-79e807205863\" (UID: \"2436e674-f08d-40d2-9c86-79e807205863\") " Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.608259 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ceph" (OuterVolumeSpecName: "ceph") pod "2436e674-f08d-40d2-9c86-79e807205863" (UID: "2436e674-f08d-40d2-9c86-79e807205863"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.608634 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2436e674-f08d-40d2-9c86-79e807205863-kube-api-access-lv7hk" (OuterVolumeSpecName: "kube-api-access-lv7hk") pod "2436e674-f08d-40d2-9c86-79e807205863" (UID: "2436e674-f08d-40d2-9c86-79e807205863"). InnerVolumeSpecName "kube-api-access-lv7hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.613649 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2436e674-f08d-40d2-9c86-79e807205863" (UID: "2436e674-f08d-40d2-9c86-79e807205863"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.648503 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-inventory" (OuterVolumeSpecName: "inventory") pod "2436e674-f08d-40d2-9c86-79e807205863" (UID: "2436e674-f08d-40d2-9c86-79e807205863"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.652067 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "2436e674-f08d-40d2-9c86-79e807205863" (UID: "2436e674-f08d-40d2-9c86-79e807205863"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.653969 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2436e674-f08d-40d2-9c86-79e807205863" (UID: "2436e674-f08d-40d2-9c86-79e807205863"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.703643 4944 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.703679 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.703693 4944 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.703704 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv7hk\" (UniqueName: \"kubernetes.io/projected/2436e674-f08d-40d2-9c86-79e807205863-kube-api-access-lv7hk\") on node \"crc\" DevicePath \"\"" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.703716 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.703728 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2436e674-f08d-40d2-9c86-79e807205863-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.900043 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" event={"ID":"2436e674-f08d-40d2-9c86-79e807205863","Type":"ContainerDied","Data":"23815a4a5da5bdd83b2eb57a2547d0ca83a7555c69287b7f38e2dbc5719109c3"} Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.900346 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23815a4a5da5bdd83b2eb57a2547d0ca83a7555c69287b7f38e2dbc5719109c3" Nov 24 11:02:07 crc kubenswrapper[4944]: I1124 11:02:07.900157 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2nlwg" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.022761 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-r796f"] Nov 24 11:02:08 crc kubenswrapper[4944]: E1124 11:02:08.023280 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72f27aa3-7622-4e53-bef1-89f30315ceb3" containerName="keystone-cron" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.023303 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="72f27aa3-7622-4e53-bef1-89f30315ceb3" containerName="keystone-cron" Nov 24 11:02:08 crc kubenswrapper[4944]: E1124 11:02:08.023365 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2436e674-f08d-40d2-9c86-79e807205863" containerName="libvirt-openstack-openstack-cell1" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.023375 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2436e674-f08d-40d2-9c86-79e807205863" containerName="libvirt-openstack-openstack-cell1" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.023618 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2436e674-f08d-40d2-9c86-79e807205863" containerName="libvirt-openstack-openstack-cell1" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.023647 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="72f27aa3-7622-4e53-bef1-89f30315ceb3" containerName="keystone-cron" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.024702 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.029384 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.029397 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.029441 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.029670 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.029967 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.030242 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.030571 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.033109 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-r796f"] Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.215599 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.215820 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.215932 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.215975 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.216035 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jgmb\" (UniqueName: \"kubernetes.io/projected/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-kube-api-access-9jgmb\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.216091 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-inventory\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.216118 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.216182 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.216260 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.216405 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.216458 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ceph\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.318698 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.318800 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ceph\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.318899 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.318927 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.318970 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.318994 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.319030 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jgmb\" (UniqueName: \"kubernetes.io/projected/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-kube-api-access-9jgmb\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.319069 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-inventory\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.319087 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.319148 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.319183 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.319607 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.320847 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.322907 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.325739 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.333727 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.334186 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-inventory\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.335031 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.343698 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ceph\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.351706 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.358607 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.363938 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jgmb\" (UniqueName: \"kubernetes.io/projected/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-kube-api-access-9jgmb\") pod \"nova-cell1-openstack-openstack-cell1-r796f\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:08 crc kubenswrapper[4944]: I1124 11:02:08.644694 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:02:09 crc kubenswrapper[4944]: I1124 11:02:09.164699 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-r796f"] Nov 24 11:02:09 crc kubenswrapper[4944]: I1124 11:02:09.928438 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" event={"ID":"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef","Type":"ContainerStarted","Data":"5c02a6881976ede98b7fbee68eb268a068d57d2117a1eb6bf0e31355d78cb132"} Nov 24 11:02:10 crc kubenswrapper[4944]: I1124 11:02:10.944959 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" event={"ID":"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef","Type":"ContainerStarted","Data":"9bd3b9e11ec235bdf786bcb2e78ba809173784d2be2dbb1ec355b393836ff5b4"} Nov 24 11:02:10 crc kubenswrapper[4944]: I1124 11:02:10.976633 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" podStartSLOduration=3.445409406 podStartE2EDuration="3.976612655s" podCreationTimestamp="2025-11-24 11:02:07 +0000 UTC" firstStartedPulling="2025-11-24 11:02:09.190295323 +0000 UTC m=+7789.724735785" lastFinishedPulling="2025-11-24 11:02:09.721498542 +0000 UTC m=+7790.255939034" observedRunningTime="2025-11-24 11:02:10.968759845 +0000 UTC m=+7791.503200327" watchObservedRunningTime="2025-11-24 11:02:10.976612655 +0000 UTC m=+7791.511053117" Nov 24 11:02:19 crc kubenswrapper[4944]: I1124 11:02:19.277232 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:02:19 crc kubenswrapper[4944]: E1124 11:02:19.278020 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:02:34 crc kubenswrapper[4944]: I1124 11:02:34.277095 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:02:34 crc kubenswrapper[4944]: E1124 11:02:34.277974 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.356029 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b5t7n"] Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.360498 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.385570 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b5t7n"] Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.455583 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-utilities\") pod \"redhat-operators-b5t7n\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.455776 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-catalog-content\") pod \"redhat-operators-b5t7n\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.455818 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4xqp\" (UniqueName: \"kubernetes.io/projected/f10c65ba-ff73-400f-81e0-fd9370cd1a59-kube-api-access-p4xqp\") pod \"redhat-operators-b5t7n\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.558241 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-catalog-content\") pod \"redhat-operators-b5t7n\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.558747 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4xqp\" (UniqueName: \"kubernetes.io/projected/f10c65ba-ff73-400f-81e0-fd9370cd1a59-kube-api-access-p4xqp\") pod \"redhat-operators-b5t7n\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.558689 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-catalog-content\") pod \"redhat-operators-b5t7n\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.559253 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-utilities\") pod \"redhat-operators-b5t7n\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.559534 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-utilities\") pod \"redhat-operators-b5t7n\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.587028 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4xqp\" (UniqueName: \"kubernetes.io/projected/f10c65ba-ff73-400f-81e0-fd9370cd1a59-kube-api-access-p4xqp\") pod \"redhat-operators-b5t7n\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:42 crc kubenswrapper[4944]: I1124 11:02:42.701310 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:43 crc kubenswrapper[4944]: I1124 11:02:43.198492 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b5t7n"] Nov 24 11:02:43 crc kubenswrapper[4944]: I1124 11:02:43.282187 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5t7n" event={"ID":"f10c65ba-ff73-400f-81e0-fd9370cd1a59","Type":"ContainerStarted","Data":"fb0e32c69d4d71a5c9cb0fe922f88a422a44f995f899fc0e453946a83dea261e"} Nov 24 11:02:44 crc kubenswrapper[4944]: I1124 11:02:44.292683 4944 generic.go:334] "Generic (PLEG): container finished" podID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerID="39e74fa52b5a5718cfee498d010dbd4e5d4f44add51a9129ac04584aca253422" exitCode=0 Nov 24 11:02:44 crc kubenswrapper[4944]: I1124 11:02:44.292730 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5t7n" event={"ID":"f10c65ba-ff73-400f-81e0-fd9370cd1a59","Type":"ContainerDied","Data":"39e74fa52b5a5718cfee498d010dbd4e5d4f44add51a9129ac04584aca253422"} Nov 24 11:02:45 crc kubenswrapper[4944]: I1124 11:02:45.306362 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5t7n" event={"ID":"f10c65ba-ff73-400f-81e0-fd9370cd1a59","Type":"ContainerStarted","Data":"eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996"} Nov 24 11:02:48 crc kubenswrapper[4944]: I1124 11:02:48.433976 4944 generic.go:334] "Generic (PLEG): container finished" podID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerID="eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996" exitCode=0 Nov 24 11:02:48 crc kubenswrapper[4944]: I1124 11:02:48.434719 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5t7n" event={"ID":"f10c65ba-ff73-400f-81e0-fd9370cd1a59","Type":"ContainerDied","Data":"eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996"} Nov 24 11:02:49 crc kubenswrapper[4944]: I1124 11:02:49.278665 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:02:49 crc kubenswrapper[4944]: E1124 11:02:49.279689 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:02:49 crc kubenswrapper[4944]: I1124 11:02:49.446744 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5t7n" event={"ID":"f10c65ba-ff73-400f-81e0-fd9370cd1a59","Type":"ContainerStarted","Data":"688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f"} Nov 24 11:02:49 crc kubenswrapper[4944]: I1124 11:02:49.469182 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b5t7n" podStartSLOduration=2.909630658 podStartE2EDuration="7.469157768s" podCreationTimestamp="2025-11-24 11:02:42 +0000 UTC" firstStartedPulling="2025-11-24 11:02:44.294410213 +0000 UTC m=+7824.828850695" lastFinishedPulling="2025-11-24 11:02:48.853937343 +0000 UTC m=+7829.388377805" observedRunningTime="2025-11-24 11:02:49.466118703 +0000 UTC m=+7830.000559205" watchObservedRunningTime="2025-11-24 11:02:49.469157768 +0000 UTC m=+7830.003598270" Nov 24 11:02:52 crc kubenswrapper[4944]: I1124 11:02:52.702793 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:52 crc kubenswrapper[4944]: I1124 11:02:52.703848 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:02:53 crc kubenswrapper[4944]: I1124 11:02:53.764103 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b5t7n" podUID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerName="registry-server" probeResult="failure" output=< Nov 24 11:02:53 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 11:02:53 crc kubenswrapper[4944]: > Nov 24 11:03:02 crc kubenswrapper[4944]: I1124 11:03:02.759143 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:03:02 crc kubenswrapper[4944]: I1124 11:03:02.809101 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:03:02 crc kubenswrapper[4944]: I1124 11:03:02.994401 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b5t7n"] Nov 24 11:03:04 crc kubenswrapper[4944]: I1124 11:03:04.277266 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:03:04 crc kubenswrapper[4944]: E1124 11:03:04.277961 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:03:04 crc kubenswrapper[4944]: I1124 11:03:04.594987 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b5t7n" podUID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerName="registry-server" containerID="cri-o://688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f" gracePeriod=2 Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.084448 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.198609 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-catalog-content\") pod \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.198913 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4xqp\" (UniqueName: \"kubernetes.io/projected/f10c65ba-ff73-400f-81e0-fd9370cd1a59-kube-api-access-p4xqp\") pod \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.198969 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-utilities\") pod \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\" (UID: \"f10c65ba-ff73-400f-81e0-fd9370cd1a59\") " Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.199975 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-utilities" (OuterVolumeSpecName: "utilities") pod "f10c65ba-ff73-400f-81e0-fd9370cd1a59" (UID: "f10c65ba-ff73-400f-81e0-fd9370cd1a59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.204416 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f10c65ba-ff73-400f-81e0-fd9370cd1a59-kube-api-access-p4xqp" (OuterVolumeSpecName: "kube-api-access-p4xqp") pod "f10c65ba-ff73-400f-81e0-fd9370cd1a59" (UID: "f10c65ba-ff73-400f-81e0-fd9370cd1a59"). InnerVolumeSpecName "kube-api-access-p4xqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.285363 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f10c65ba-ff73-400f-81e0-fd9370cd1a59" (UID: "f10c65ba-ff73-400f-81e0-fd9370cd1a59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.302281 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4xqp\" (UniqueName: \"kubernetes.io/projected/f10c65ba-ff73-400f-81e0-fd9370cd1a59-kube-api-access-p4xqp\") on node \"crc\" DevicePath \"\"" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.302369 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.302388 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f10c65ba-ff73-400f-81e0-fd9370cd1a59-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.610275 4944 generic.go:334] "Generic (PLEG): container finished" podID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerID="688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f" exitCode=0 Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.610350 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5t7n" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.610380 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5t7n" event={"ID":"f10c65ba-ff73-400f-81e0-fd9370cd1a59","Type":"ContainerDied","Data":"688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f"} Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.610687 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5t7n" event={"ID":"f10c65ba-ff73-400f-81e0-fd9370cd1a59","Type":"ContainerDied","Data":"fb0e32c69d4d71a5c9cb0fe922f88a422a44f995f899fc0e453946a83dea261e"} Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.610722 4944 scope.go:117] "RemoveContainer" containerID="688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.649453 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b5t7n"] Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.652606 4944 scope.go:117] "RemoveContainer" containerID="eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.658858 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b5t7n"] Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.683932 4944 scope.go:117] "RemoveContainer" containerID="39e74fa52b5a5718cfee498d010dbd4e5d4f44add51a9129ac04584aca253422" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.749337 4944 scope.go:117] "RemoveContainer" containerID="688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f" Nov 24 11:03:05 crc kubenswrapper[4944]: E1124 11:03:05.750536 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f\": container with ID starting with 688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f not found: ID does not exist" containerID="688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.750599 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f"} err="failed to get container status \"688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f\": rpc error: code = NotFound desc = could not find container \"688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f\": container with ID starting with 688a0e217db8f5304519ddc96da75fefabab1d9f21f93d9a62ea8b5f3802b22f not found: ID does not exist" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.750641 4944 scope.go:117] "RemoveContainer" containerID="eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996" Nov 24 11:03:05 crc kubenswrapper[4944]: E1124 11:03:05.751087 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996\": container with ID starting with eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996 not found: ID does not exist" containerID="eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.751217 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996"} err="failed to get container status \"eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996\": rpc error: code = NotFound desc = could not find container \"eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996\": container with ID starting with eb02d92bc4bdfc614f87d4f3b7fd513bf4bf620bca9c15a1f7fe7778a89bb996 not found: ID does not exist" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.751332 4944 scope.go:117] "RemoveContainer" containerID="39e74fa52b5a5718cfee498d010dbd4e5d4f44add51a9129ac04584aca253422" Nov 24 11:03:05 crc kubenswrapper[4944]: E1124 11:03:05.751752 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39e74fa52b5a5718cfee498d010dbd4e5d4f44add51a9129ac04584aca253422\": container with ID starting with 39e74fa52b5a5718cfee498d010dbd4e5d4f44add51a9129ac04584aca253422 not found: ID does not exist" containerID="39e74fa52b5a5718cfee498d010dbd4e5d4f44add51a9129ac04584aca253422" Nov 24 11:03:05 crc kubenswrapper[4944]: I1124 11:03:05.751794 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39e74fa52b5a5718cfee498d010dbd4e5d4f44add51a9129ac04584aca253422"} err="failed to get container status \"39e74fa52b5a5718cfee498d010dbd4e5d4f44add51a9129ac04584aca253422\": rpc error: code = NotFound desc = could not find container \"39e74fa52b5a5718cfee498d010dbd4e5d4f44add51a9129ac04584aca253422\": container with ID starting with 39e74fa52b5a5718cfee498d010dbd4e5d4f44add51a9129ac04584aca253422 not found: ID does not exist" Nov 24 11:03:06 crc kubenswrapper[4944]: I1124 11:03:06.292506 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" path="/var/lib/kubelet/pods/f10c65ba-ff73-400f-81e0-fd9370cd1a59/volumes" Nov 24 11:03:16 crc kubenswrapper[4944]: I1124 11:03:16.279145 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:03:16 crc kubenswrapper[4944]: E1124 11:03:16.280771 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:03:29 crc kubenswrapper[4944]: I1124 11:03:29.276872 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:03:29 crc kubenswrapper[4944]: E1124 11:03:29.277692 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:03:44 crc kubenswrapper[4944]: I1124 11:03:44.278812 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:03:44 crc kubenswrapper[4944]: E1124 11:03:44.279516 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:03:57 crc kubenswrapper[4944]: I1124 11:03:57.277738 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:03:57 crc kubenswrapper[4944]: E1124 11:03:57.278854 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:04:11 crc kubenswrapper[4944]: I1124 11:04:11.277821 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:04:11 crc kubenswrapper[4944]: E1124 11:04:11.278627 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:04:23 crc kubenswrapper[4944]: I1124 11:04:23.278426 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:04:23 crc kubenswrapper[4944]: E1124 11:04:23.279285 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:04:37 crc kubenswrapper[4944]: I1124 11:04:37.277044 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:04:37 crc kubenswrapper[4944]: E1124 11:04:37.277645 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:04:49 crc kubenswrapper[4944]: I1124 11:04:49.276643 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:04:49 crc kubenswrapper[4944]: E1124 11:04:49.277276 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:05:00 crc kubenswrapper[4944]: I1124 11:05:00.301218 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:05:00 crc kubenswrapper[4944]: E1124 11:05:00.302352 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:05:07 crc kubenswrapper[4944]: E1124 11:05:07.218916 4944 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod453bfabf_cd8c_46b1_b036_a1e16f1cf2ef.slice/crio-conmon-9bd3b9e11ec235bdf786bcb2e78ba809173784d2be2dbb1ec355b393836ff5b4.scope\": RecentStats: unable to find data in memory cache]" Nov 24 11:05:08 crc kubenswrapper[4944]: I1124 11:05:08.023726 4944 generic.go:334] "Generic (PLEG): container finished" podID="453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" containerID="9bd3b9e11ec235bdf786bcb2e78ba809173784d2be2dbb1ec355b393836ff5b4" exitCode=0 Nov 24 11:05:08 crc kubenswrapper[4944]: I1124 11:05:08.023765 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" event={"ID":"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef","Type":"ContainerDied","Data":"9bd3b9e11ec235bdf786bcb2e78ba809173784d2be2dbb1ec355b393836ff5b4"} Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.637089 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.787703 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jgmb\" (UniqueName: \"kubernetes.io/projected/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-kube-api-access-9jgmb\") pod \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.787950 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-0\") pod \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.788008 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-inventory\") pod \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.788035 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-1\") pod \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.788191 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ssh-key\") pod \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.788251 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-1\") pod \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.788297 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-0\") pod \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.788408 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-0\") pod \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.788440 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ceph\") pod \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.788483 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-1\") pod \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.788525 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-combined-ca-bundle\") pod \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\" (UID: \"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef\") " Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.795243 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-kube-api-access-9jgmb" (OuterVolumeSpecName: "kube-api-access-9jgmb") pod "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" (UID: "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef"). InnerVolumeSpecName "kube-api-access-9jgmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.795461 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" (UID: "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.801368 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ceph" (OuterVolumeSpecName: "ceph") pod "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" (UID: "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.816308 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" (UID: "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.822129 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-inventory" (OuterVolumeSpecName: "inventory") pod "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" (UID: "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.822933 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" (UID: "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.823935 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" (UID: "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.826020 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" (UID: "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.829321 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" (UID: "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.832423 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" (UID: "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.847483 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" (UID: "453bfabf-cd8c-46b1-b036-a1e16f1cf2ef"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.891483 4944 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.891742 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.891961 4944 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.892024 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.892086 4944 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.892148 4944 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.892196 4944 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.892244 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.892290 4944 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.892337 4944 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:05:09 crc kubenswrapper[4944]: I1124 11:05:09.892417 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jgmb\" (UniqueName: \"kubernetes.io/projected/453bfabf-cd8c-46b1-b036-a1e16f1cf2ef-kube-api-access-9jgmb\") on node \"crc\" DevicePath \"\"" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.044235 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" event={"ID":"453bfabf-cd8c-46b1-b036-a1e16f1cf2ef","Type":"ContainerDied","Data":"5c02a6881976ede98b7fbee68eb268a068d57d2117a1eb6bf0e31355d78cb132"} Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.044274 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c02a6881976ede98b7fbee68eb268a068d57d2117a1eb6bf0e31355d78cb132" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.044351 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-r796f" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.155774 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-299ff"] Nov 24 11:05:10 crc kubenswrapper[4944]: E1124 11:05:10.156320 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerName="registry-server" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.156342 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerName="registry-server" Nov 24 11:05:10 crc kubenswrapper[4944]: E1124 11:05:10.156353 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerName="extract-utilities" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.156362 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerName="extract-utilities" Nov 24 11:05:10 crc kubenswrapper[4944]: E1124 11:05:10.156384 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerName="extract-content" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.156392 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerName="extract-content" Nov 24 11:05:10 crc kubenswrapper[4944]: E1124 11:05:10.156417 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" containerName="nova-cell1-openstack-openstack-cell1" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.156427 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" containerName="nova-cell1-openstack-openstack-cell1" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.156674 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="453bfabf-cd8c-46b1-b036-a1e16f1cf2ef" containerName="nova-cell1-openstack-openstack-cell1" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.156703 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f10c65ba-ff73-400f-81e0-fd9370cd1a59" containerName="registry-server" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.157597 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.163373 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.163519 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.163606 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.164385 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.164528 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.177561 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-299ff"] Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.303146 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.303216 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-inventory\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.303246 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.303393 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.303649 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.303770 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb2c9\" (UniqueName: \"kubernetes.io/projected/9a9fdd29-f803-41a5-96eb-dc882bedc95a-kube-api-access-tb2c9\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.303826 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceph\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.303904 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ssh-key\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.405304 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ssh-key\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.405629 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.405679 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-inventory\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.405703 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.405746 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.405793 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.405849 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb2c9\" (UniqueName: \"kubernetes.io/projected/9a9fdd29-f803-41a5-96eb-dc882bedc95a-kube-api-access-tb2c9\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.405881 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceph\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.409251 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.409278 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.409605 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.410613 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ssh-key\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.410874 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.411519 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-inventory\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.412560 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceph\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.424024 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb2c9\" (UniqueName: \"kubernetes.io/projected/9a9fdd29-f803-41a5-96eb-dc882bedc95a-kube-api-access-tb2c9\") pod \"telemetry-openstack-openstack-cell1-299ff\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:10 crc kubenswrapper[4944]: I1124 11:05:10.481898 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:05:11 crc kubenswrapper[4944]: I1124 11:05:11.061286 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-299ff"] Nov 24 11:05:11 crc kubenswrapper[4944]: I1124 11:05:11.068071 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 11:05:12 crc kubenswrapper[4944]: I1124 11:05:12.078754 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-299ff" event={"ID":"9a9fdd29-f803-41a5-96eb-dc882bedc95a","Type":"ContainerStarted","Data":"30831f1b91a0de5c113130377aa2a5b988c8ed82827eb9ff3e5e42d11c7501ec"} Nov 24 11:05:12 crc kubenswrapper[4944]: I1124 11:05:12.079110 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-299ff" event={"ID":"9a9fdd29-f803-41a5-96eb-dc882bedc95a","Type":"ContainerStarted","Data":"c9726915121ab4385b76e642bdbbd6117f12f600359e09c15cba2ac4afb86736"} Nov 24 11:05:12 crc kubenswrapper[4944]: I1124 11:05:12.101965 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-299ff" podStartSLOduration=1.506209896 podStartE2EDuration="2.101946653s" podCreationTimestamp="2025-11-24 11:05:10 +0000 UTC" firstStartedPulling="2025-11-24 11:05:11.067861383 +0000 UTC m=+7971.602301845" lastFinishedPulling="2025-11-24 11:05:11.66359814 +0000 UTC m=+7972.198038602" observedRunningTime="2025-11-24 11:05:12.094326996 +0000 UTC m=+7972.628767458" watchObservedRunningTime="2025-11-24 11:05:12.101946653 +0000 UTC m=+7972.636387115" Nov 24 11:05:15 crc kubenswrapper[4944]: I1124 11:05:15.276962 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:05:15 crc kubenswrapper[4944]: E1124 11:05:15.278698 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:05:28 crc kubenswrapper[4944]: I1124 11:05:28.277714 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:05:28 crc kubenswrapper[4944]: E1124 11:05:28.278673 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:05:39 crc kubenswrapper[4944]: I1124 11:05:39.277713 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:05:39 crc kubenswrapper[4944]: E1124 11:05:39.278777 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:05:51 crc kubenswrapper[4944]: I1124 11:05:51.277897 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:05:51 crc kubenswrapper[4944]: E1124 11:05:51.278648 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:06:05 crc kubenswrapper[4944]: I1124 11:06:05.278368 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:06:05 crc kubenswrapper[4944]: E1124 11:06:05.280412 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:06:19 crc kubenswrapper[4944]: I1124 11:06:19.276593 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:06:19 crc kubenswrapper[4944]: E1124 11:06:19.277488 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:06:34 crc kubenswrapper[4944]: I1124 11:06:34.277291 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:06:34 crc kubenswrapper[4944]: E1124 11:06:34.277967 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:06:47 crc kubenswrapper[4944]: I1124 11:06:47.276856 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:06:47 crc kubenswrapper[4944]: E1124 11:06:47.277652 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:06:59 crc kubenswrapper[4944]: I1124 11:06:59.277145 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:07:00 crc kubenswrapper[4944]: I1124 11:07:00.144472 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"b17321821c14c72225af1578f9d68638570a79b91360efb5de60c020d3bd3aa4"} Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.395025 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qmgxx"] Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.398988 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.417472 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmgxx"] Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.573325 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6r7x\" (UniqueName: \"kubernetes.io/projected/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-kube-api-access-f6r7x\") pod \"redhat-marketplace-qmgxx\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.573509 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-catalog-content\") pod \"redhat-marketplace-qmgxx\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.573676 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-utilities\") pod \"redhat-marketplace-qmgxx\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.676094 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-catalog-content\") pod \"redhat-marketplace-qmgxx\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.676193 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-utilities\") pod \"redhat-marketplace-qmgxx\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.676351 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6r7x\" (UniqueName: \"kubernetes.io/projected/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-kube-api-access-f6r7x\") pod \"redhat-marketplace-qmgxx\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.676659 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-catalog-content\") pod \"redhat-marketplace-qmgxx\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.676729 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-utilities\") pod \"redhat-marketplace-qmgxx\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.696968 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6r7x\" (UniqueName: \"kubernetes.io/projected/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-kube-api-access-f6r7x\") pod \"redhat-marketplace-qmgxx\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:02 crc kubenswrapper[4944]: I1124 11:09:02.719229 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:03 crc kubenswrapper[4944]: I1124 11:09:03.200926 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmgxx"] Nov 24 11:09:03 crc kubenswrapper[4944]: I1124 11:09:03.349309 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmgxx" event={"ID":"47f703cf-082d-4dcb-936c-4ac21ac7a3d4","Type":"ContainerStarted","Data":"90be3a69d16a753f89038810a1289284ff9f5fec302b6c130727a0c221cda92e"} Nov 24 11:09:04 crc kubenswrapper[4944]: I1124 11:09:04.364672 4944 generic.go:334] "Generic (PLEG): container finished" podID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" containerID="51cf4ba32a2209943aff65e7396fed1272c368422c6623e16584934432524733" exitCode=0 Nov 24 11:09:04 crc kubenswrapper[4944]: I1124 11:09:04.364915 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmgxx" event={"ID":"47f703cf-082d-4dcb-936c-4ac21ac7a3d4","Type":"ContainerDied","Data":"51cf4ba32a2209943aff65e7396fed1272c368422c6623e16584934432524733"} Nov 24 11:09:05 crc kubenswrapper[4944]: I1124 11:09:05.374795 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmgxx" event={"ID":"47f703cf-082d-4dcb-936c-4ac21ac7a3d4","Type":"ContainerStarted","Data":"0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c"} Nov 24 11:09:06 crc kubenswrapper[4944]: I1124 11:09:06.390036 4944 generic.go:334] "Generic (PLEG): container finished" podID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" containerID="0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c" exitCode=0 Nov 24 11:09:06 crc kubenswrapper[4944]: I1124 11:09:06.390087 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmgxx" event={"ID":"47f703cf-082d-4dcb-936c-4ac21ac7a3d4","Type":"ContainerDied","Data":"0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c"} Nov 24 11:09:06 crc kubenswrapper[4944]: I1124 11:09:06.395423 4944 generic.go:334] "Generic (PLEG): container finished" podID="9a9fdd29-f803-41a5-96eb-dc882bedc95a" containerID="30831f1b91a0de5c113130377aa2a5b988c8ed82827eb9ff3e5e42d11c7501ec" exitCode=0 Nov 24 11:09:06 crc kubenswrapper[4944]: I1124 11:09:06.395464 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-299ff" event={"ID":"9a9fdd29-f803-41a5-96eb-dc882bedc95a","Type":"ContainerDied","Data":"30831f1b91a0de5c113130377aa2a5b988c8ed82827eb9ff3e5e42d11c7501ec"} Nov 24 11:09:07 crc kubenswrapper[4944]: I1124 11:09:07.407795 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmgxx" event={"ID":"47f703cf-082d-4dcb-936c-4ac21ac7a3d4","Type":"ContainerStarted","Data":"dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f"} Nov 24 11:09:07 crc kubenswrapper[4944]: I1124 11:09:07.432375 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qmgxx" podStartSLOduration=2.915258858 podStartE2EDuration="5.432356313s" podCreationTimestamp="2025-11-24 11:09:02 +0000 UTC" firstStartedPulling="2025-11-24 11:09:04.367422087 +0000 UTC m=+8204.901862549" lastFinishedPulling="2025-11-24 11:09:06.884519542 +0000 UTC m=+8207.418960004" observedRunningTime="2025-11-24 11:09:07.427398838 +0000 UTC m=+8207.961839320" watchObservedRunningTime="2025-11-24 11:09:07.432356313 +0000 UTC m=+8207.966796775" Nov 24 11:09:07 crc kubenswrapper[4944]: I1124 11:09:07.935692 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.006585 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-inventory\") pod \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.006690 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-telemetry-combined-ca-bundle\") pod \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.006731 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ssh-key\") pod \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.006796 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb2c9\" (UniqueName: \"kubernetes.io/projected/9a9fdd29-f803-41a5-96eb-dc882bedc95a-kube-api-access-tb2c9\") pod \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.006850 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-2\") pod \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.007779 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceph\") pod \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.007907 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-0\") pod \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.007940 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-1\") pod \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\" (UID: \"9a9fdd29-f803-41a5-96eb-dc882bedc95a\") " Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.016209 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceph" (OuterVolumeSpecName: "ceph") pod "9a9fdd29-f803-41a5-96eb-dc882bedc95a" (UID: "9a9fdd29-f803-41a5-96eb-dc882bedc95a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.016563 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "9a9fdd29-f803-41a5-96eb-dc882bedc95a" (UID: "9a9fdd29-f803-41a5-96eb-dc882bedc95a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.017096 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a9fdd29-f803-41a5-96eb-dc882bedc95a-kube-api-access-tb2c9" (OuterVolumeSpecName: "kube-api-access-tb2c9") pod "9a9fdd29-f803-41a5-96eb-dc882bedc95a" (UID: "9a9fdd29-f803-41a5-96eb-dc882bedc95a"). InnerVolumeSpecName "kube-api-access-tb2c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.038987 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-inventory" (OuterVolumeSpecName: "inventory") pod "9a9fdd29-f803-41a5-96eb-dc882bedc95a" (UID: "9a9fdd29-f803-41a5-96eb-dc882bedc95a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.039540 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "9a9fdd29-f803-41a5-96eb-dc882bedc95a" (UID: "9a9fdd29-f803-41a5-96eb-dc882bedc95a"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.042781 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9a9fdd29-f803-41a5-96eb-dc882bedc95a" (UID: "9a9fdd29-f803-41a5-96eb-dc882bedc95a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.050565 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "9a9fdd29-f803-41a5-96eb-dc882bedc95a" (UID: "9a9fdd29-f803-41a5-96eb-dc882bedc95a"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.053357 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "9a9fdd29-f803-41a5-96eb-dc882bedc95a" (UID: "9a9fdd29-f803-41a5-96eb-dc882bedc95a"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.110342 4944 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.110377 4944 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.110393 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.110405 4944 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.110418 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.110430 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb2c9\" (UniqueName: \"kubernetes.io/projected/9a9fdd29-f803-41a5-96eb-dc882bedc95a-kube-api-access-tb2c9\") on node \"crc\" DevicePath \"\"" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.110442 4944 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.110452 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a9fdd29-f803-41a5-96eb-dc882bedc95a-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.421360 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-299ff" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.421583 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-299ff" event={"ID":"9a9fdd29-f803-41a5-96eb-dc882bedc95a","Type":"ContainerDied","Data":"c9726915121ab4385b76e642bdbbd6117f12f600359e09c15cba2ac4afb86736"} Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.422098 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9726915121ab4385b76e642bdbbd6117f12f600359e09c15cba2ac4afb86736" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.533588 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-k5pgj"] Nov 24 11:09:08 crc kubenswrapper[4944]: E1124 11:09:08.533996 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a9fdd29-f803-41a5-96eb-dc882bedc95a" containerName="telemetry-openstack-openstack-cell1" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.534013 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a9fdd29-f803-41a5-96eb-dc882bedc95a" containerName="telemetry-openstack-openstack-cell1" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.534242 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a9fdd29-f803-41a5-96eb-dc882bedc95a" containerName="telemetry-openstack-openstack-cell1" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.534956 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.539830 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.539831 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.539865 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.540252 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.553438 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.553637 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-k5pgj"] Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.720827 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.720891 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.721587 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg5hb\" (UniqueName: \"kubernetes.io/projected/20c0c984-e49b-42f6-b3ee-385aa90befc5-kube-api-access-dg5hb\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.722945 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.723017 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.723413 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.826274 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.826876 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.827214 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.827447 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.827666 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg5hb\" (UniqueName: \"kubernetes.io/projected/20c0c984-e49b-42f6-b3ee-385aa90befc5-kube-api-access-dg5hb\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.827945 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.831533 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.832878 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.840882 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.843277 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.844187 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg5hb\" (UniqueName: \"kubernetes.io/projected/20c0c984-e49b-42f6-b3ee-385aa90befc5-kube-api-access-dg5hb\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.845381 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-k5pgj\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:08 crc kubenswrapper[4944]: I1124 11:09:08.854147 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:09:09 crc kubenswrapper[4944]: I1124 11:09:09.391257 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-k5pgj"] Nov 24 11:09:09 crc kubenswrapper[4944]: W1124 11:09:09.395483 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20c0c984_e49b_42f6_b3ee_385aa90befc5.slice/crio-67da0f7710e1a5f6ce30bfe252e3c9b438598fd7ee9044052d24ee2b96ef8aae WatchSource:0}: Error finding container 67da0f7710e1a5f6ce30bfe252e3c9b438598fd7ee9044052d24ee2b96ef8aae: Status 404 returned error can't find the container with id 67da0f7710e1a5f6ce30bfe252e3c9b438598fd7ee9044052d24ee2b96ef8aae Nov 24 11:09:09 crc kubenswrapper[4944]: I1124 11:09:09.432641 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" event={"ID":"20c0c984-e49b-42f6-b3ee-385aa90befc5","Type":"ContainerStarted","Data":"67da0f7710e1a5f6ce30bfe252e3c9b438598fd7ee9044052d24ee2b96ef8aae"} Nov 24 11:09:10 crc kubenswrapper[4944]: I1124 11:09:10.444962 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" event={"ID":"20c0c984-e49b-42f6-b3ee-385aa90befc5","Type":"ContainerStarted","Data":"8d8626061004febb746c4330cc38cd6edfad146d9465cfc3cb827265de9e0085"} Nov 24 11:09:10 crc kubenswrapper[4944]: I1124 11:09:10.472914 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" podStartSLOduration=1.996215557 podStartE2EDuration="2.472892167s" podCreationTimestamp="2025-11-24 11:09:08 +0000 UTC" firstStartedPulling="2025-11-24 11:09:09.398827249 +0000 UTC m=+8209.933267711" lastFinishedPulling="2025-11-24 11:09:09.875503849 +0000 UTC m=+8210.409944321" observedRunningTime="2025-11-24 11:09:10.46146232 +0000 UTC m=+8210.995902822" watchObservedRunningTime="2025-11-24 11:09:10.472892167 +0000 UTC m=+8211.007332649" Nov 24 11:09:12 crc kubenswrapper[4944]: I1124 11:09:12.719723 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:12 crc kubenswrapper[4944]: I1124 11:09:12.719974 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:12 crc kubenswrapper[4944]: I1124 11:09:12.788028 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:13 crc kubenswrapper[4944]: I1124 11:09:13.533158 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:13 crc kubenswrapper[4944]: I1124 11:09:13.595471 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmgxx"] Nov 24 11:09:15 crc kubenswrapper[4944]: I1124 11:09:15.492413 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qmgxx" podUID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" containerName="registry-server" containerID="cri-o://dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f" gracePeriod=2 Nov 24 11:09:15 crc kubenswrapper[4944]: I1124 11:09:15.963381 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.087649 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-catalog-content\") pod \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.087733 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-utilities\") pod \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.087949 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6r7x\" (UniqueName: \"kubernetes.io/projected/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-kube-api-access-f6r7x\") pod \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\" (UID: \"47f703cf-082d-4dcb-936c-4ac21ac7a3d4\") " Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.088782 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-utilities" (OuterVolumeSpecName: "utilities") pod "47f703cf-082d-4dcb-936c-4ac21ac7a3d4" (UID: "47f703cf-082d-4dcb-936c-4ac21ac7a3d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.094824 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-kube-api-access-f6r7x" (OuterVolumeSpecName: "kube-api-access-f6r7x") pod "47f703cf-082d-4dcb-936c-4ac21ac7a3d4" (UID: "47f703cf-082d-4dcb-936c-4ac21ac7a3d4"). InnerVolumeSpecName "kube-api-access-f6r7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.105743 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47f703cf-082d-4dcb-936c-4ac21ac7a3d4" (UID: "47f703cf-082d-4dcb-936c-4ac21ac7a3d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.191309 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.191346 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.191359 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6r7x\" (UniqueName: \"kubernetes.io/projected/47f703cf-082d-4dcb-936c-4ac21ac7a3d4-kube-api-access-f6r7x\") on node \"crc\" DevicePath \"\"" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.505745 4944 generic.go:334] "Generic (PLEG): container finished" podID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" containerID="dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f" exitCode=0 Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.505805 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmgxx" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.505829 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmgxx" event={"ID":"47f703cf-082d-4dcb-936c-4ac21ac7a3d4","Type":"ContainerDied","Data":"dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f"} Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.505864 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmgxx" event={"ID":"47f703cf-082d-4dcb-936c-4ac21ac7a3d4","Type":"ContainerDied","Data":"90be3a69d16a753f89038810a1289284ff9f5fec302b6c130727a0c221cda92e"} Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.505885 4944 scope.go:117] "RemoveContainer" containerID="dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.534431 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmgxx"] Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.543601 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmgxx"] Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.550225 4944 scope.go:117] "RemoveContainer" containerID="0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.582583 4944 scope.go:117] "RemoveContainer" containerID="51cf4ba32a2209943aff65e7396fed1272c368422c6623e16584934432524733" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.625412 4944 scope.go:117] "RemoveContainer" containerID="dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f" Nov 24 11:09:16 crc kubenswrapper[4944]: E1124 11:09:16.625925 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f\": container with ID starting with dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f not found: ID does not exist" containerID="dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.625971 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f"} err="failed to get container status \"dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f\": rpc error: code = NotFound desc = could not find container \"dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f\": container with ID starting with dfd19c6f1e95aa91a000ad56fc59880c720fec4d58182dea6e3d02fde8f47a4f not found: ID does not exist" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.625996 4944 scope.go:117] "RemoveContainer" containerID="0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c" Nov 24 11:09:16 crc kubenswrapper[4944]: E1124 11:09:16.626419 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c\": container with ID starting with 0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c not found: ID does not exist" containerID="0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.626453 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c"} err="failed to get container status \"0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c\": rpc error: code = NotFound desc = could not find container \"0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c\": container with ID starting with 0db2de1a3f6bb4a67c879272755c33837e4cccf7b961c47a5bf17fe0c8012d4c not found: ID does not exist" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.626478 4944 scope.go:117] "RemoveContainer" containerID="51cf4ba32a2209943aff65e7396fed1272c368422c6623e16584934432524733" Nov 24 11:09:16 crc kubenswrapper[4944]: E1124 11:09:16.626703 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51cf4ba32a2209943aff65e7396fed1272c368422c6623e16584934432524733\": container with ID starting with 51cf4ba32a2209943aff65e7396fed1272c368422c6623e16584934432524733 not found: ID does not exist" containerID="51cf4ba32a2209943aff65e7396fed1272c368422c6623e16584934432524733" Nov 24 11:09:16 crc kubenswrapper[4944]: I1124 11:09:16.626727 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51cf4ba32a2209943aff65e7396fed1272c368422c6623e16584934432524733"} err="failed to get container status \"51cf4ba32a2209943aff65e7396fed1272c368422c6623e16584934432524733\": rpc error: code = NotFound desc = could not find container \"51cf4ba32a2209943aff65e7396fed1272c368422c6623e16584934432524733\": container with ID starting with 51cf4ba32a2209943aff65e7396fed1272c368422c6623e16584934432524733 not found: ID does not exist" Nov 24 11:09:18 crc kubenswrapper[4944]: I1124 11:09:18.295684 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" path="/var/lib/kubelet/pods/47f703cf-082d-4dcb-936c-4ac21ac7a3d4/volumes" Nov 24 11:09:23 crc kubenswrapper[4944]: I1124 11:09:23.548702 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:09:23 crc kubenswrapper[4944]: I1124 11:09:23.549183 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:09:53 crc kubenswrapper[4944]: I1124 11:09:53.548403 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:09:53 crc kubenswrapper[4944]: I1124 11:09:53.548998 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:10:23 crc kubenswrapper[4944]: I1124 11:10:23.549128 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:10:23 crc kubenswrapper[4944]: I1124 11:10:23.549827 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:10:23 crc kubenswrapper[4944]: I1124 11:10:23.549894 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 11:10:23 crc kubenswrapper[4944]: I1124 11:10:23.551024 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b17321821c14c72225af1578f9d68638570a79b91360efb5de60c020d3bd3aa4"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 11:10:23 crc kubenswrapper[4944]: I1124 11:10:23.551107 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://b17321821c14c72225af1578f9d68638570a79b91360efb5de60c020d3bd3aa4" gracePeriod=600 Nov 24 11:10:24 crc kubenswrapper[4944]: I1124 11:10:24.167727 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="b17321821c14c72225af1578f9d68638570a79b91360efb5de60c020d3bd3aa4" exitCode=0 Nov 24 11:10:24 crc kubenswrapper[4944]: I1124 11:10:24.167802 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"b17321821c14c72225af1578f9d68638570a79b91360efb5de60c020d3bd3aa4"} Nov 24 11:10:24 crc kubenswrapper[4944]: I1124 11:10:24.168428 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d"} Nov 24 11:10:24 crc kubenswrapper[4944]: I1124 11:10:24.168460 4944 scope.go:117] "RemoveContainer" containerID="5a699efb3f2fdc0756dcfea20a9add56ccd42f96179de19deae70ff66476a5c2" Nov 24 11:11:56 crc kubenswrapper[4944]: I1124 11:11:56.978390 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-24r2s"] Nov 24 11:11:56 crc kubenswrapper[4944]: E1124 11:11:56.980375 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" containerName="registry-server" Nov 24 11:11:56 crc kubenswrapper[4944]: I1124 11:11:56.980400 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" containerName="registry-server" Nov 24 11:11:56 crc kubenswrapper[4944]: E1124 11:11:56.980427 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" containerName="extract-content" Nov 24 11:11:56 crc kubenswrapper[4944]: I1124 11:11:56.980436 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" containerName="extract-content" Nov 24 11:11:56 crc kubenswrapper[4944]: E1124 11:11:56.980452 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" containerName="extract-utilities" Nov 24 11:11:56 crc kubenswrapper[4944]: I1124 11:11:56.980463 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" containerName="extract-utilities" Nov 24 11:11:56 crc kubenswrapper[4944]: I1124 11:11:56.981132 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="47f703cf-082d-4dcb-936c-4ac21ac7a3d4" containerName="registry-server" Nov 24 11:11:56 crc kubenswrapper[4944]: I1124 11:11:56.990305 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.000470 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24r2s"] Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.097767 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2zxg\" (UniqueName: \"kubernetes.io/projected/49b3ab41-c219-435e-a804-7b945d5b3e9e-kube-api-access-l2zxg\") pod \"certified-operators-24r2s\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.098295 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-utilities\") pod \"certified-operators-24r2s\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.098473 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-catalog-content\") pod \"certified-operators-24r2s\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.200208 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-catalog-content\") pod \"certified-operators-24r2s\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.200540 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2zxg\" (UniqueName: \"kubernetes.io/projected/49b3ab41-c219-435e-a804-7b945d5b3e9e-kube-api-access-l2zxg\") pod \"certified-operators-24r2s\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.200712 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-utilities\") pod \"certified-operators-24r2s\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.200761 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-catalog-content\") pod \"certified-operators-24r2s\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.201173 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-utilities\") pod \"certified-operators-24r2s\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.231949 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2zxg\" (UniqueName: \"kubernetes.io/projected/49b3ab41-c219-435e-a804-7b945d5b3e9e-kube-api-access-l2zxg\") pod \"certified-operators-24r2s\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.323867 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:11:57 crc kubenswrapper[4944]: I1124 11:11:57.938060 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24r2s"] Nov 24 11:11:58 crc kubenswrapper[4944]: I1124 11:11:58.148021 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24r2s" event={"ID":"49b3ab41-c219-435e-a804-7b945d5b3e9e","Type":"ContainerStarted","Data":"16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550"} Nov 24 11:11:58 crc kubenswrapper[4944]: I1124 11:11:58.148352 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24r2s" event={"ID":"49b3ab41-c219-435e-a804-7b945d5b3e9e","Type":"ContainerStarted","Data":"b5c46dbdae4b93e08cedb5bf85bf3c50b4503b615b21c78a199d8ad59132bfb9"} Nov 24 11:11:58 crc kubenswrapper[4944]: I1124 11:11:58.150612 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 11:11:59 crc kubenswrapper[4944]: I1124 11:11:59.158920 4944 generic.go:334] "Generic (PLEG): container finished" podID="49b3ab41-c219-435e-a804-7b945d5b3e9e" containerID="16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550" exitCode=0 Nov 24 11:11:59 crc kubenswrapper[4944]: I1124 11:11:59.159002 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24r2s" event={"ID":"49b3ab41-c219-435e-a804-7b945d5b3e9e","Type":"ContainerDied","Data":"16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550"} Nov 24 11:11:59 crc kubenswrapper[4944]: I1124 11:11:59.159328 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24r2s" event={"ID":"49b3ab41-c219-435e-a804-7b945d5b3e9e","Type":"ContainerStarted","Data":"b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48"} Nov 24 11:12:00 crc kubenswrapper[4944]: I1124 11:12:00.172749 4944 generic.go:334] "Generic (PLEG): container finished" podID="49b3ab41-c219-435e-a804-7b945d5b3e9e" containerID="b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48" exitCode=0 Nov 24 11:12:00 crc kubenswrapper[4944]: I1124 11:12:00.172856 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24r2s" event={"ID":"49b3ab41-c219-435e-a804-7b945d5b3e9e","Type":"ContainerDied","Data":"b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48"} Nov 24 11:12:01 crc kubenswrapper[4944]: I1124 11:12:01.185808 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24r2s" event={"ID":"49b3ab41-c219-435e-a804-7b945d5b3e9e","Type":"ContainerStarted","Data":"6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace"} Nov 24 11:12:01 crc kubenswrapper[4944]: I1124 11:12:01.213128 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-24r2s" podStartSLOduration=2.558517433 podStartE2EDuration="5.21309917s" podCreationTimestamp="2025-11-24 11:11:56 +0000 UTC" firstStartedPulling="2025-11-24 11:11:58.150166366 +0000 UTC m=+8378.684606828" lastFinishedPulling="2025-11-24 11:12:00.804748093 +0000 UTC m=+8381.339188565" observedRunningTime="2025-11-24 11:12:01.206751432 +0000 UTC m=+8381.741191934" watchObservedRunningTime="2025-11-24 11:12:01.21309917 +0000 UTC m=+8381.747539672" Nov 24 11:12:07 crc kubenswrapper[4944]: I1124 11:12:07.325291 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:12:07 crc kubenswrapper[4944]: I1124 11:12:07.325806 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:12:07 crc kubenswrapper[4944]: I1124 11:12:07.389018 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:12:08 crc kubenswrapper[4944]: I1124 11:12:08.301436 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:12:08 crc kubenswrapper[4944]: I1124 11:12:08.354953 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24r2s"] Nov 24 11:12:10 crc kubenswrapper[4944]: I1124 11:12:10.291261 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-24r2s" podUID="49b3ab41-c219-435e-a804-7b945d5b3e9e" containerName="registry-server" containerID="cri-o://6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace" gracePeriod=2 Nov 24 11:12:10 crc kubenswrapper[4944]: I1124 11:12:10.739845 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:12:10 crc kubenswrapper[4944]: I1124 11:12:10.786542 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-utilities\") pod \"49b3ab41-c219-435e-a804-7b945d5b3e9e\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " Nov 24 11:12:10 crc kubenswrapper[4944]: I1124 11:12:10.786903 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-catalog-content\") pod \"49b3ab41-c219-435e-a804-7b945d5b3e9e\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " Nov 24 11:12:10 crc kubenswrapper[4944]: I1124 11:12:10.786986 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2zxg\" (UniqueName: \"kubernetes.io/projected/49b3ab41-c219-435e-a804-7b945d5b3e9e-kube-api-access-l2zxg\") pod \"49b3ab41-c219-435e-a804-7b945d5b3e9e\" (UID: \"49b3ab41-c219-435e-a804-7b945d5b3e9e\") " Nov 24 11:12:10 crc kubenswrapper[4944]: I1124 11:12:10.788202 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-utilities" (OuterVolumeSpecName: "utilities") pod "49b3ab41-c219-435e-a804-7b945d5b3e9e" (UID: "49b3ab41-c219-435e-a804-7b945d5b3e9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:12:10 crc kubenswrapper[4944]: I1124 11:12:10.793698 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49b3ab41-c219-435e-a804-7b945d5b3e9e-kube-api-access-l2zxg" (OuterVolumeSpecName: "kube-api-access-l2zxg") pod "49b3ab41-c219-435e-a804-7b945d5b3e9e" (UID: "49b3ab41-c219-435e-a804-7b945d5b3e9e"). InnerVolumeSpecName "kube-api-access-l2zxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:12:10 crc kubenswrapper[4944]: I1124 11:12:10.831569 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49b3ab41-c219-435e-a804-7b945d5b3e9e" (UID: "49b3ab41-c219-435e-a804-7b945d5b3e9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:12:10 crc kubenswrapper[4944]: I1124 11:12:10.888931 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:12:10 crc kubenswrapper[4944]: I1124 11:12:10.888975 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3ab41-c219-435e-a804-7b945d5b3e9e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:12:10 crc kubenswrapper[4944]: I1124 11:12:10.888988 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2zxg\" (UniqueName: \"kubernetes.io/projected/49b3ab41-c219-435e-a804-7b945d5b3e9e-kube-api-access-l2zxg\") on node \"crc\" DevicePath \"\"" Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.307588 4944 generic.go:334] "Generic (PLEG): container finished" podID="49b3ab41-c219-435e-a804-7b945d5b3e9e" containerID="6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace" exitCode=0 Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.307685 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24r2s" Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.307695 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24r2s" event={"ID":"49b3ab41-c219-435e-a804-7b945d5b3e9e","Type":"ContainerDied","Data":"6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace"} Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.307997 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24r2s" event={"ID":"49b3ab41-c219-435e-a804-7b945d5b3e9e","Type":"ContainerDied","Data":"b5c46dbdae4b93e08cedb5bf85bf3c50b4503b615b21c78a199d8ad59132bfb9"} Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.308016 4944 scope.go:117] "RemoveContainer" containerID="6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace" Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.352709 4944 scope.go:117] "RemoveContainer" containerID="b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48" Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.355465 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24r2s"] Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.370307 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-24r2s"] Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.423016 4944 scope.go:117] "RemoveContainer" containerID="16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550" Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.448735 4944 scope.go:117] "RemoveContainer" containerID="6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace" Nov 24 11:12:11 crc kubenswrapper[4944]: E1124 11:12:11.449095 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace\": container with ID starting with 6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace not found: ID does not exist" containerID="6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace" Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.449123 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace"} err="failed to get container status \"6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace\": rpc error: code = NotFound desc = could not find container \"6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace\": container with ID starting with 6f097c4ef254cbd38580913db5f9a46d338669c0664fbe6abfde4cba991fdace not found: ID does not exist" Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.449144 4944 scope.go:117] "RemoveContainer" containerID="b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48" Nov 24 11:12:11 crc kubenswrapper[4944]: E1124 11:12:11.449458 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48\": container with ID starting with b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48 not found: ID does not exist" containerID="b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48" Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.449478 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48"} err="failed to get container status \"b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48\": rpc error: code = NotFound desc = could not find container \"b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48\": container with ID starting with b83c12621fa51cda6ea358c29899e1b4cfa9866a40409a1ac1c025aad1047e48 not found: ID does not exist" Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.449492 4944 scope.go:117] "RemoveContainer" containerID="16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550" Nov 24 11:12:11 crc kubenswrapper[4944]: E1124 11:12:11.449801 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550\": container with ID starting with 16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550 not found: ID does not exist" containerID="16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550" Nov 24 11:12:11 crc kubenswrapper[4944]: I1124 11:12:11.449822 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550"} err="failed to get container status \"16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550\": rpc error: code = NotFound desc = could not find container \"16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550\": container with ID starting with 16cc02d99e2499d2dde0a57357be94c9baf8f87fe5b1c2557fcd11b4c9ccb550 not found: ID does not exist" Nov 24 11:12:12 crc kubenswrapper[4944]: I1124 11:12:12.288482 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49b3ab41-c219-435e-a804-7b945d5b3e9e" path="/var/lib/kubelet/pods/49b3ab41-c219-435e-a804-7b945d5b3e9e/volumes" Nov 24 11:12:23 crc kubenswrapper[4944]: I1124 11:12:23.549122 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:12:23 crc kubenswrapper[4944]: I1124 11:12:23.549650 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:12:53 crc kubenswrapper[4944]: I1124 11:12:53.548740 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:12:53 crc kubenswrapper[4944]: I1124 11:12:53.550565 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.218476 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-47n25"] Nov 24 11:13:13 crc kubenswrapper[4944]: E1124 11:13:13.219639 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b3ab41-c219-435e-a804-7b945d5b3e9e" containerName="extract-content" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.219657 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b3ab41-c219-435e-a804-7b945d5b3e9e" containerName="extract-content" Nov 24 11:13:13 crc kubenswrapper[4944]: E1124 11:13:13.219677 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b3ab41-c219-435e-a804-7b945d5b3e9e" containerName="extract-utilities" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.219686 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b3ab41-c219-435e-a804-7b945d5b3e9e" containerName="extract-utilities" Nov 24 11:13:13 crc kubenswrapper[4944]: E1124 11:13:13.219717 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b3ab41-c219-435e-a804-7b945d5b3e9e" containerName="registry-server" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.219725 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b3ab41-c219-435e-a804-7b945d5b3e9e" containerName="registry-server" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.220009 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="49b3ab41-c219-435e-a804-7b945d5b3e9e" containerName="registry-server" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.224100 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.243951 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-47n25"] Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.332784 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-catalog-content\") pod \"redhat-operators-47n25\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.332842 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8rff\" (UniqueName: \"kubernetes.io/projected/c0a3babe-7afa-4fac-afac-3ac246604eb8-kube-api-access-j8rff\") pod \"redhat-operators-47n25\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.332896 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-utilities\") pod \"redhat-operators-47n25\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.435070 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-utilities\") pod \"redhat-operators-47n25\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.435318 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-catalog-content\") pod \"redhat-operators-47n25\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.435355 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8rff\" (UniqueName: \"kubernetes.io/projected/c0a3babe-7afa-4fac-afac-3ac246604eb8-kube-api-access-j8rff\") pod \"redhat-operators-47n25\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.435613 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-utilities\") pod \"redhat-operators-47n25\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.435710 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-catalog-content\") pod \"redhat-operators-47n25\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.455931 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8rff\" (UniqueName: \"kubernetes.io/projected/c0a3babe-7afa-4fac-afac-3ac246604eb8-kube-api-access-j8rff\") pod \"redhat-operators-47n25\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:13 crc kubenswrapper[4944]: I1124 11:13:13.568924 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:14 crc kubenswrapper[4944]: I1124 11:13:14.045649 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-47n25"] Nov 24 11:13:14 crc kubenswrapper[4944]: I1124 11:13:14.977107 4944 generic.go:334] "Generic (PLEG): container finished" podID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerID="c946781dccb6db0ad0b961bf935ff0fe5e12f24f929ef58134e0adbe9beb54fa" exitCode=0 Nov 24 11:13:14 crc kubenswrapper[4944]: I1124 11:13:14.977152 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47n25" event={"ID":"c0a3babe-7afa-4fac-afac-3ac246604eb8","Type":"ContainerDied","Data":"c946781dccb6db0ad0b961bf935ff0fe5e12f24f929ef58134e0adbe9beb54fa"} Nov 24 11:13:14 crc kubenswrapper[4944]: I1124 11:13:14.977382 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47n25" event={"ID":"c0a3babe-7afa-4fac-afac-3ac246604eb8","Type":"ContainerStarted","Data":"d8bdf3c2e53e9bf25906b1461acb8be4a7b28f489bebe7348e870af301d19d91"} Nov 24 11:13:15 crc kubenswrapper[4944]: I1124 11:13:15.988772 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47n25" event={"ID":"c0a3babe-7afa-4fac-afac-3ac246604eb8","Type":"ContainerStarted","Data":"440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91"} Nov 24 11:13:18 crc kubenswrapper[4944]: I1124 11:13:18.007745 4944 generic.go:334] "Generic (PLEG): container finished" podID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerID="440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91" exitCode=0 Nov 24 11:13:18 crc kubenswrapper[4944]: I1124 11:13:18.007821 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47n25" event={"ID":"c0a3babe-7afa-4fac-afac-3ac246604eb8","Type":"ContainerDied","Data":"440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91"} Nov 24 11:13:19 crc kubenswrapper[4944]: I1124 11:13:19.021411 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47n25" event={"ID":"c0a3babe-7afa-4fac-afac-3ac246604eb8","Type":"ContainerStarted","Data":"6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b"} Nov 24 11:13:19 crc kubenswrapper[4944]: I1124 11:13:19.042373 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-47n25" podStartSLOduration=2.529057506 podStartE2EDuration="6.042354028s" podCreationTimestamp="2025-11-24 11:13:13 +0000 UTC" firstStartedPulling="2025-11-24 11:13:14.978668055 +0000 UTC m=+8455.513108517" lastFinishedPulling="2025-11-24 11:13:18.491964577 +0000 UTC m=+8459.026405039" observedRunningTime="2025-11-24 11:13:19.037523917 +0000 UTC m=+8459.571964379" watchObservedRunningTime="2025-11-24 11:13:19.042354028 +0000 UTC m=+8459.576794490" Nov 24 11:13:23 crc kubenswrapper[4944]: I1124 11:13:23.548874 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:13:23 crc kubenswrapper[4944]: I1124 11:13:23.549468 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:13:23 crc kubenswrapper[4944]: I1124 11:13:23.549512 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 11:13:23 crc kubenswrapper[4944]: I1124 11:13:23.550371 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 11:13:23 crc kubenswrapper[4944]: I1124 11:13:23.550424 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" gracePeriod=600 Nov 24 11:13:23 crc kubenswrapper[4944]: I1124 11:13:23.569291 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:23 crc kubenswrapper[4944]: I1124 11:13:23.569351 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:23 crc kubenswrapper[4944]: E1124 11:13:23.679931 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:13:24 crc kubenswrapper[4944]: I1124 11:13:24.088606 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" exitCode=0 Nov 24 11:13:24 crc kubenswrapper[4944]: I1124 11:13:24.088683 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d"} Nov 24 11:13:24 crc kubenswrapper[4944]: I1124 11:13:24.089275 4944 scope.go:117] "RemoveContainer" containerID="b17321821c14c72225af1578f9d68638570a79b91360efb5de60c020d3bd3aa4" Nov 24 11:13:24 crc kubenswrapper[4944]: I1124 11:13:24.089872 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:13:24 crc kubenswrapper[4944]: E1124 11:13:24.091160 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:13:24 crc kubenswrapper[4944]: I1124 11:13:24.613174 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-47n25" podUID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerName="registry-server" probeResult="failure" output=< Nov 24 11:13:24 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 11:13:24 crc kubenswrapper[4944]: > Nov 24 11:13:33 crc kubenswrapper[4944]: I1124 11:13:33.613228 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:33 crc kubenswrapper[4944]: I1124 11:13:33.661926 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:33 crc kubenswrapper[4944]: I1124 11:13:33.857441 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-47n25"] Nov 24 11:13:35 crc kubenswrapper[4944]: I1124 11:13:35.212816 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-47n25" podUID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerName="registry-server" containerID="cri-o://6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b" gracePeriod=2 Nov 24 11:13:35 crc kubenswrapper[4944]: I1124 11:13:35.682258 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:35 crc kubenswrapper[4944]: I1124 11:13:35.778027 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-utilities\") pod \"c0a3babe-7afa-4fac-afac-3ac246604eb8\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " Nov 24 11:13:35 crc kubenswrapper[4944]: I1124 11:13:35.778249 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8rff\" (UniqueName: \"kubernetes.io/projected/c0a3babe-7afa-4fac-afac-3ac246604eb8-kube-api-access-j8rff\") pod \"c0a3babe-7afa-4fac-afac-3ac246604eb8\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " Nov 24 11:13:35 crc kubenswrapper[4944]: I1124 11:13:35.778325 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-catalog-content\") pod \"c0a3babe-7afa-4fac-afac-3ac246604eb8\" (UID: \"c0a3babe-7afa-4fac-afac-3ac246604eb8\") " Nov 24 11:13:35 crc kubenswrapper[4944]: I1124 11:13:35.778950 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-utilities" (OuterVolumeSpecName: "utilities") pod "c0a3babe-7afa-4fac-afac-3ac246604eb8" (UID: "c0a3babe-7afa-4fac-afac-3ac246604eb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:13:35 crc kubenswrapper[4944]: I1124 11:13:35.783797 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0a3babe-7afa-4fac-afac-3ac246604eb8-kube-api-access-j8rff" (OuterVolumeSpecName: "kube-api-access-j8rff") pod "c0a3babe-7afa-4fac-afac-3ac246604eb8" (UID: "c0a3babe-7afa-4fac-afac-3ac246604eb8"). InnerVolumeSpecName "kube-api-access-j8rff". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:13:35 crc kubenswrapper[4944]: I1124 11:13:35.873906 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0a3babe-7afa-4fac-afac-3ac246604eb8" (UID: "c0a3babe-7afa-4fac-afac-3ac246604eb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:13:35 crc kubenswrapper[4944]: I1124 11:13:35.880682 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8rff\" (UniqueName: \"kubernetes.io/projected/c0a3babe-7afa-4fac-afac-3ac246604eb8-kube-api-access-j8rff\") on node \"crc\" DevicePath \"\"" Nov 24 11:13:35 crc kubenswrapper[4944]: I1124 11:13:35.880725 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:13:35 crc kubenswrapper[4944]: I1124 11:13:35.880736 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0a3babe-7afa-4fac-afac-3ac246604eb8-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.224345 4944 generic.go:334] "Generic (PLEG): container finished" podID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerID="6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b" exitCode=0 Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.224386 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47n25" event={"ID":"c0a3babe-7afa-4fac-afac-3ac246604eb8","Type":"ContainerDied","Data":"6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b"} Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.224430 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-47n25" Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.224457 4944 scope.go:117] "RemoveContainer" containerID="6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b" Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.224440 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-47n25" event={"ID":"c0a3babe-7afa-4fac-afac-3ac246604eb8","Type":"ContainerDied","Data":"d8bdf3c2e53e9bf25906b1461acb8be4a7b28f489bebe7348e870af301d19d91"} Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.247262 4944 scope.go:117] "RemoveContainer" containerID="440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91" Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.261212 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-47n25"] Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.270127 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-47n25"] Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.287668 4944 scope.go:117] "RemoveContainer" containerID="c946781dccb6db0ad0b961bf935ff0fe5e12f24f929ef58134e0adbe9beb54fa" Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.292035 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0a3babe-7afa-4fac-afac-3ac246604eb8" path="/var/lib/kubelet/pods/c0a3babe-7afa-4fac-afac-3ac246604eb8/volumes" Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.328251 4944 scope.go:117] "RemoveContainer" containerID="6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b" Nov 24 11:13:36 crc kubenswrapper[4944]: E1124 11:13:36.339488 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b\": container with ID starting with 6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b not found: ID does not exist" containerID="6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b" Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.339534 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b"} err="failed to get container status \"6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b\": rpc error: code = NotFound desc = could not find container \"6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b\": container with ID starting with 6cc4307724f18e4f6caf1086bcb60139f781ea7eb5404599467649536414b35b not found: ID does not exist" Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.339558 4944 scope.go:117] "RemoveContainer" containerID="440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91" Nov 24 11:13:36 crc kubenswrapper[4944]: E1124 11:13:36.339817 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91\": container with ID starting with 440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91 not found: ID does not exist" containerID="440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91" Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.339845 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91"} err="failed to get container status \"440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91\": rpc error: code = NotFound desc = could not find container \"440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91\": container with ID starting with 440b174ca4fc57ac6546512cad4bb2a77eb36c11ca9eae28d8257ba8c45baa91 not found: ID does not exist" Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.339859 4944 scope.go:117] "RemoveContainer" containerID="c946781dccb6db0ad0b961bf935ff0fe5e12f24f929ef58134e0adbe9beb54fa" Nov 24 11:13:36 crc kubenswrapper[4944]: E1124 11:13:36.340014 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c946781dccb6db0ad0b961bf935ff0fe5e12f24f929ef58134e0adbe9beb54fa\": container with ID starting with c946781dccb6db0ad0b961bf935ff0fe5e12f24f929ef58134e0adbe9beb54fa not found: ID does not exist" containerID="c946781dccb6db0ad0b961bf935ff0fe5e12f24f929ef58134e0adbe9beb54fa" Nov 24 11:13:36 crc kubenswrapper[4944]: I1124 11:13:36.340044 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c946781dccb6db0ad0b961bf935ff0fe5e12f24f929ef58134e0adbe9beb54fa"} err="failed to get container status \"c946781dccb6db0ad0b961bf935ff0fe5e12f24f929ef58134e0adbe9beb54fa\": rpc error: code = NotFound desc = could not find container \"c946781dccb6db0ad0b961bf935ff0fe5e12f24f929ef58134e0adbe9beb54fa\": container with ID starting with c946781dccb6db0ad0b961bf935ff0fe5e12f24f929ef58134e0adbe9beb54fa not found: ID does not exist" Nov 24 11:13:37 crc kubenswrapper[4944]: I1124 11:13:37.277316 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:13:37 crc kubenswrapper[4944]: E1124 11:13:37.277666 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.068472 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xlfpb"] Nov 24 11:13:41 crc kubenswrapper[4944]: E1124 11:13:41.070215 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerName="extract-content" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.070300 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerName="extract-content" Nov 24 11:13:41 crc kubenswrapper[4944]: E1124 11:13:41.070372 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerName="registry-server" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.070426 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerName="registry-server" Nov 24 11:13:41 crc kubenswrapper[4944]: E1124 11:13:41.070504 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerName="extract-utilities" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.070563 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerName="extract-utilities" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.070841 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0a3babe-7afa-4fac-afac-3ac246604eb8" containerName="registry-server" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.072459 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.080096 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xlfpb"] Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.203927 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-catalog-content\") pod \"community-operators-xlfpb\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.203980 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-utilities\") pod \"community-operators-xlfpb\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.204332 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cwqr\" (UniqueName: \"kubernetes.io/projected/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-kube-api-access-2cwqr\") pod \"community-operators-xlfpb\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.306251 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-catalog-content\") pod \"community-operators-xlfpb\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.306633 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-utilities\") pod \"community-operators-xlfpb\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.306763 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-catalog-content\") pod \"community-operators-xlfpb\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.307011 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-utilities\") pod \"community-operators-xlfpb\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.307267 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cwqr\" (UniqueName: \"kubernetes.io/projected/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-kube-api-access-2cwqr\") pod \"community-operators-xlfpb\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.348868 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cwqr\" (UniqueName: \"kubernetes.io/projected/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-kube-api-access-2cwqr\") pod \"community-operators-xlfpb\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.407713 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:41 crc kubenswrapper[4944]: I1124 11:13:41.955000 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xlfpb"] Nov 24 11:13:42 crc kubenswrapper[4944]: I1124 11:13:42.289003 4944 generic.go:334] "Generic (PLEG): container finished" podID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" containerID="88db3f3f9f3cee19d15586a36f995d7a1e8cd810d3af5dc003f26570316c054b" exitCode=0 Nov 24 11:13:42 crc kubenswrapper[4944]: I1124 11:13:42.289039 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xlfpb" event={"ID":"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19","Type":"ContainerDied","Data":"88db3f3f9f3cee19d15586a36f995d7a1e8cd810d3af5dc003f26570316c054b"} Nov 24 11:13:42 crc kubenswrapper[4944]: I1124 11:13:42.289076 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xlfpb" event={"ID":"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19","Type":"ContainerStarted","Data":"1d20e60e21706a14ec71b5b7e55606cef85d7f29ccb8434a138a074237af5d06"} Nov 24 11:13:43 crc kubenswrapper[4944]: I1124 11:13:43.302039 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xlfpb" event={"ID":"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19","Type":"ContainerStarted","Data":"d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577"} Nov 24 11:13:44 crc kubenswrapper[4944]: I1124 11:13:44.316085 4944 generic.go:334] "Generic (PLEG): container finished" podID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" containerID="d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577" exitCode=0 Nov 24 11:13:44 crc kubenswrapper[4944]: I1124 11:13:44.316188 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xlfpb" event={"ID":"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19","Type":"ContainerDied","Data":"d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577"} Nov 24 11:13:45 crc kubenswrapper[4944]: I1124 11:13:45.335973 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xlfpb" event={"ID":"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19","Type":"ContainerStarted","Data":"fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825"} Nov 24 11:13:45 crc kubenswrapper[4944]: I1124 11:13:45.360298 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xlfpb" podStartSLOduration=1.862515184 podStartE2EDuration="4.360278595s" podCreationTimestamp="2025-11-24 11:13:41 +0000 UTC" firstStartedPulling="2025-11-24 11:13:42.290709264 +0000 UTC m=+8482.825149726" lastFinishedPulling="2025-11-24 11:13:44.788472675 +0000 UTC m=+8485.322913137" observedRunningTime="2025-11-24 11:13:45.357719685 +0000 UTC m=+8485.892160157" watchObservedRunningTime="2025-11-24 11:13:45.360278595 +0000 UTC m=+8485.894719057" Nov 24 11:13:51 crc kubenswrapper[4944]: I1124 11:13:51.408231 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:51 crc kubenswrapper[4944]: I1124 11:13:51.408863 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:51 crc kubenswrapper[4944]: I1124 11:13:51.464773 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:52 crc kubenswrapper[4944]: I1124 11:13:52.278450 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:13:52 crc kubenswrapper[4944]: E1124 11:13:52.279144 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:13:52 crc kubenswrapper[4944]: I1124 11:13:52.476723 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:54 crc kubenswrapper[4944]: I1124 11:13:54.655781 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xlfpb"] Nov 24 11:13:55 crc kubenswrapper[4944]: I1124 11:13:55.456270 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xlfpb" podUID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" containerName="registry-server" containerID="cri-o://fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825" gracePeriod=2 Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.000037 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.129859 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-utilities\") pod \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.130012 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cwqr\" (UniqueName: \"kubernetes.io/projected/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-kube-api-access-2cwqr\") pod \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.130158 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-catalog-content\") pod \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\" (UID: \"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19\") " Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.130957 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-utilities" (OuterVolumeSpecName: "utilities") pod "3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" (UID: "3dcf2a6c-04e5-4d39-9bb6-88446ee92d19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.140930 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-kube-api-access-2cwqr" (OuterVolumeSpecName: "kube-api-access-2cwqr") pod "3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" (UID: "3dcf2a6c-04e5-4d39-9bb6-88446ee92d19"). InnerVolumeSpecName "kube-api-access-2cwqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.187208 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" (UID: "3dcf2a6c-04e5-4d39-9bb6-88446ee92d19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.234032 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.234083 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cwqr\" (UniqueName: \"kubernetes.io/projected/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-kube-api-access-2cwqr\") on node \"crc\" DevicePath \"\"" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.234094 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.477206 4944 generic.go:334] "Generic (PLEG): container finished" podID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" containerID="fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825" exitCode=0 Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.477257 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xlfpb" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.477257 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xlfpb" event={"ID":"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19","Type":"ContainerDied","Data":"fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825"} Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.477479 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xlfpb" event={"ID":"3dcf2a6c-04e5-4d39-9bb6-88446ee92d19","Type":"ContainerDied","Data":"1d20e60e21706a14ec71b5b7e55606cef85d7f29ccb8434a138a074237af5d06"} Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.477530 4944 scope.go:117] "RemoveContainer" containerID="fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.510436 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xlfpb"] Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.522282 4944 scope.go:117] "RemoveContainer" containerID="d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.526494 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xlfpb"] Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.560980 4944 scope.go:117] "RemoveContainer" containerID="88db3f3f9f3cee19d15586a36f995d7a1e8cd810d3af5dc003f26570316c054b" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.600918 4944 scope.go:117] "RemoveContainer" containerID="fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825" Nov 24 11:13:56 crc kubenswrapper[4944]: E1124 11:13:56.601472 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825\": container with ID starting with fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825 not found: ID does not exist" containerID="fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.601515 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825"} err="failed to get container status \"fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825\": rpc error: code = NotFound desc = could not find container \"fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825\": container with ID starting with fa8e6c05fd487ea7e31576662503f8d232e492a02082aff222e3cd5e89bf1825 not found: ID does not exist" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.601549 4944 scope.go:117] "RemoveContainer" containerID="d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577" Nov 24 11:13:56 crc kubenswrapper[4944]: E1124 11:13:56.602093 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577\": container with ID starting with d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577 not found: ID does not exist" containerID="d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.602149 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577"} err="failed to get container status \"d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577\": rpc error: code = NotFound desc = could not find container \"d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577\": container with ID starting with d0bad08ad5e71833e3f75d5cda590fa396d403b5771ceae155954b209d1d1577 not found: ID does not exist" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.602192 4944 scope.go:117] "RemoveContainer" containerID="88db3f3f9f3cee19d15586a36f995d7a1e8cd810d3af5dc003f26570316c054b" Nov 24 11:13:56 crc kubenswrapper[4944]: E1124 11:13:56.602597 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88db3f3f9f3cee19d15586a36f995d7a1e8cd810d3af5dc003f26570316c054b\": container with ID starting with 88db3f3f9f3cee19d15586a36f995d7a1e8cd810d3af5dc003f26570316c054b not found: ID does not exist" containerID="88db3f3f9f3cee19d15586a36f995d7a1e8cd810d3af5dc003f26570316c054b" Nov 24 11:13:56 crc kubenswrapper[4944]: I1124 11:13:56.602628 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88db3f3f9f3cee19d15586a36f995d7a1e8cd810d3af5dc003f26570316c054b"} err="failed to get container status \"88db3f3f9f3cee19d15586a36f995d7a1e8cd810d3af5dc003f26570316c054b\": rpc error: code = NotFound desc = could not find container \"88db3f3f9f3cee19d15586a36f995d7a1e8cd810d3af5dc003f26570316c054b\": container with ID starting with 88db3f3f9f3cee19d15586a36f995d7a1e8cd810d3af5dc003f26570316c054b not found: ID does not exist" Nov 24 11:13:58 crc kubenswrapper[4944]: I1124 11:13:58.294690 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" path="/var/lib/kubelet/pods/3dcf2a6c-04e5-4d39-9bb6-88446ee92d19/volumes" Nov 24 11:14:06 crc kubenswrapper[4944]: I1124 11:14:06.276841 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:14:06 crc kubenswrapper[4944]: E1124 11:14:06.277953 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:14:18 crc kubenswrapper[4944]: I1124 11:14:18.277587 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:14:18 crc kubenswrapper[4944]: E1124 11:14:18.278313 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:14:32 crc kubenswrapper[4944]: I1124 11:14:32.459136 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:14:32 crc kubenswrapper[4944]: E1124 11:14:32.459978 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:14:43 crc kubenswrapper[4944]: I1124 11:14:43.277194 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:14:43 crc kubenswrapper[4944]: E1124 11:14:43.299856 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:14:57 crc kubenswrapper[4944]: I1124 11:14:57.276908 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:14:57 crc kubenswrapper[4944]: E1124 11:14:57.277795 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.152562 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw"] Nov 24 11:15:00 crc kubenswrapper[4944]: E1124 11:15:00.153554 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" containerName="extract-content" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.153575 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" containerName="extract-content" Nov 24 11:15:00 crc kubenswrapper[4944]: E1124 11:15:00.153611 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" containerName="extract-utilities" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.153619 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" containerName="extract-utilities" Nov 24 11:15:00 crc kubenswrapper[4944]: E1124 11:15:00.153654 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" containerName="registry-server" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.153662 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" containerName="registry-server" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.153955 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dcf2a6c-04e5-4d39-9bb6-88446ee92d19" containerName="registry-server" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.155103 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.157526 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.158136 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.167474 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw"] Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.344452 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a85ac49-8904-4650-961a-418e5144c293-secret-volume\") pod \"collect-profiles-29399715-5jrvw\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.344509 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a85ac49-8904-4650-961a-418e5144c293-config-volume\") pod \"collect-profiles-29399715-5jrvw\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.344609 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn948\" (UniqueName: \"kubernetes.io/projected/1a85ac49-8904-4650-961a-418e5144c293-kube-api-access-mn948\") pod \"collect-profiles-29399715-5jrvw\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.446168 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn948\" (UniqueName: \"kubernetes.io/projected/1a85ac49-8904-4650-961a-418e5144c293-kube-api-access-mn948\") pod \"collect-profiles-29399715-5jrvw\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.447021 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a85ac49-8904-4650-961a-418e5144c293-secret-volume\") pod \"collect-profiles-29399715-5jrvw\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.447221 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a85ac49-8904-4650-961a-418e5144c293-config-volume\") pod \"collect-profiles-29399715-5jrvw\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.448261 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a85ac49-8904-4650-961a-418e5144c293-config-volume\") pod \"collect-profiles-29399715-5jrvw\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.455200 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a85ac49-8904-4650-961a-418e5144c293-secret-volume\") pod \"collect-profiles-29399715-5jrvw\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.470732 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn948\" (UniqueName: \"kubernetes.io/projected/1a85ac49-8904-4650-961a-418e5144c293-kube-api-access-mn948\") pod \"collect-profiles-29399715-5jrvw\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.483646 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:00 crc kubenswrapper[4944]: I1124 11:15:00.926767 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw"] Nov 24 11:15:00 crc kubenswrapper[4944]: W1124 11:15:00.938958 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a85ac49_8904_4650_961a_418e5144c293.slice/crio-b278a13cdc5f174cdb9d8f8d5af564755302b5a1a065e9cefe052ace6b9e8e84 WatchSource:0}: Error finding container b278a13cdc5f174cdb9d8f8d5af564755302b5a1a065e9cefe052ace6b9e8e84: Status 404 returned error can't find the container with id b278a13cdc5f174cdb9d8f8d5af564755302b5a1a065e9cefe052ace6b9e8e84 Nov 24 11:15:01 crc kubenswrapper[4944]: I1124 11:15:01.075147 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" event={"ID":"1a85ac49-8904-4650-961a-418e5144c293","Type":"ContainerStarted","Data":"b278a13cdc5f174cdb9d8f8d5af564755302b5a1a065e9cefe052ace6b9e8e84"} Nov 24 11:15:02 crc kubenswrapper[4944]: I1124 11:15:02.087778 4944 generic.go:334] "Generic (PLEG): container finished" podID="1a85ac49-8904-4650-961a-418e5144c293" containerID="6eea4cdd8195916673e01d660f6d0bafaad94cbdd95a94bd23b43c72a14796a1" exitCode=0 Nov 24 11:15:02 crc kubenswrapper[4944]: I1124 11:15:02.087935 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" event={"ID":"1a85ac49-8904-4650-961a-418e5144c293","Type":"ContainerDied","Data":"6eea4cdd8195916673e01d660f6d0bafaad94cbdd95a94bd23b43c72a14796a1"} Nov 24 11:15:03 crc kubenswrapper[4944]: I1124 11:15:03.461361 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:03 crc kubenswrapper[4944]: I1124 11:15:03.518974 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a85ac49-8904-4650-961a-418e5144c293-secret-volume\") pod \"1a85ac49-8904-4650-961a-418e5144c293\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " Nov 24 11:15:03 crc kubenswrapper[4944]: I1124 11:15:03.519030 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a85ac49-8904-4650-961a-418e5144c293-config-volume\") pod \"1a85ac49-8904-4650-961a-418e5144c293\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " Nov 24 11:15:03 crc kubenswrapper[4944]: I1124 11:15:03.519175 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn948\" (UniqueName: \"kubernetes.io/projected/1a85ac49-8904-4650-961a-418e5144c293-kube-api-access-mn948\") pod \"1a85ac49-8904-4650-961a-418e5144c293\" (UID: \"1a85ac49-8904-4650-961a-418e5144c293\") " Nov 24 11:15:03 crc kubenswrapper[4944]: I1124 11:15:03.519907 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a85ac49-8904-4650-961a-418e5144c293-config-volume" (OuterVolumeSpecName: "config-volume") pod "1a85ac49-8904-4650-961a-418e5144c293" (UID: "1a85ac49-8904-4650-961a-418e5144c293"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 11:15:03 crc kubenswrapper[4944]: I1124 11:15:03.526234 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a85ac49-8904-4650-961a-418e5144c293-kube-api-access-mn948" (OuterVolumeSpecName: "kube-api-access-mn948") pod "1a85ac49-8904-4650-961a-418e5144c293" (UID: "1a85ac49-8904-4650-961a-418e5144c293"). InnerVolumeSpecName "kube-api-access-mn948". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:15:03 crc kubenswrapper[4944]: I1124 11:15:03.529240 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a85ac49-8904-4650-961a-418e5144c293-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1a85ac49-8904-4650-961a-418e5144c293" (UID: "1a85ac49-8904-4650-961a-418e5144c293"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:15:03 crc kubenswrapper[4944]: I1124 11:15:03.622496 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn948\" (UniqueName: \"kubernetes.io/projected/1a85ac49-8904-4650-961a-418e5144c293-kube-api-access-mn948\") on node \"crc\" DevicePath \"\"" Nov 24 11:15:03 crc kubenswrapper[4944]: I1124 11:15:03.622535 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a85ac49-8904-4650-961a-418e5144c293-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 11:15:03 crc kubenswrapper[4944]: I1124 11:15:03.622552 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a85ac49-8904-4650-961a-418e5144c293-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 11:15:04 crc kubenswrapper[4944]: I1124 11:15:04.108224 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" event={"ID":"1a85ac49-8904-4650-961a-418e5144c293","Type":"ContainerDied","Data":"b278a13cdc5f174cdb9d8f8d5af564755302b5a1a065e9cefe052ace6b9e8e84"} Nov 24 11:15:04 crc kubenswrapper[4944]: I1124 11:15:04.108269 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b278a13cdc5f174cdb9d8f8d5af564755302b5a1a065e9cefe052ace6b9e8e84" Nov 24 11:15:04 crc kubenswrapper[4944]: I1124 11:15:04.108438 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399715-5jrvw" Nov 24 11:15:04 crc kubenswrapper[4944]: I1124 11:15:04.538691 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm"] Nov 24 11:15:04 crc kubenswrapper[4944]: I1124 11:15:04.549101 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399670-4lwmm"] Nov 24 11:15:06 crc kubenswrapper[4944]: I1124 11:15:06.289795 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac6a4cc0-e137-401b-92c0-3835c028d5fa" path="/var/lib/kubelet/pods/ac6a4cc0-e137-401b-92c0-3835c028d5fa/volumes" Nov 24 11:15:08 crc kubenswrapper[4944]: I1124 11:15:08.166412 4944 generic.go:334] "Generic (PLEG): container finished" podID="20c0c984-e49b-42f6-b3ee-385aa90befc5" containerID="8d8626061004febb746c4330cc38cd6edfad146d9465cfc3cb827265de9e0085" exitCode=0 Nov 24 11:15:08 crc kubenswrapper[4944]: I1124 11:15:08.166741 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" event={"ID":"20c0c984-e49b-42f6-b3ee-385aa90befc5","Type":"ContainerDied","Data":"8d8626061004febb746c4330cc38cd6edfad146d9465cfc3cb827265de9e0085"} Nov 24 11:15:09 crc kubenswrapper[4944]: I1124 11:15:09.277137 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:15:09 crc kubenswrapper[4944]: E1124 11:15:09.277665 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:15:09 crc kubenswrapper[4944]: I1124 11:15:09.842291 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:15:09 crc kubenswrapper[4944]: I1124 11:15:09.995952 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-combined-ca-bundle\") pod \"20c0c984-e49b-42f6-b3ee-385aa90befc5\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " Nov 24 11:15:09 crc kubenswrapper[4944]: I1124 11:15:09.995998 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ssh-key\") pod \"20c0c984-e49b-42f6-b3ee-385aa90befc5\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " Nov 24 11:15:09 crc kubenswrapper[4944]: I1124 11:15:09.996025 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg5hb\" (UniqueName: \"kubernetes.io/projected/20c0c984-e49b-42f6-b3ee-385aa90befc5-kube-api-access-dg5hb\") pod \"20c0c984-e49b-42f6-b3ee-385aa90befc5\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " Nov 24 11:15:09 crc kubenswrapper[4944]: I1124 11:15:09.996184 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-agent-neutron-config-0\") pod \"20c0c984-e49b-42f6-b3ee-385aa90befc5\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " Nov 24 11:15:09 crc kubenswrapper[4944]: I1124 11:15:09.996302 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ceph\") pod \"20c0c984-e49b-42f6-b3ee-385aa90befc5\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " Nov 24 11:15:09 crc kubenswrapper[4944]: I1124 11:15:09.996355 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-inventory\") pod \"20c0c984-e49b-42f6-b3ee-385aa90befc5\" (UID: \"20c0c984-e49b-42f6-b3ee-385aa90befc5\") " Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.003465 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ceph" (OuterVolumeSpecName: "ceph") pod "20c0c984-e49b-42f6-b3ee-385aa90befc5" (UID: "20c0c984-e49b-42f6-b3ee-385aa90befc5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.003681 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "20c0c984-e49b-42f6-b3ee-385aa90befc5" (UID: "20c0c984-e49b-42f6-b3ee-385aa90befc5"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.007227 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20c0c984-e49b-42f6-b3ee-385aa90befc5-kube-api-access-dg5hb" (OuterVolumeSpecName: "kube-api-access-dg5hb") pod "20c0c984-e49b-42f6-b3ee-385aa90befc5" (UID: "20c0c984-e49b-42f6-b3ee-385aa90befc5"). InnerVolumeSpecName "kube-api-access-dg5hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.034216 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "20c0c984-e49b-42f6-b3ee-385aa90befc5" (UID: "20c0c984-e49b-42f6-b3ee-385aa90befc5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.034724 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "20c0c984-e49b-42f6-b3ee-385aa90befc5" (UID: "20c0c984-e49b-42f6-b3ee-385aa90befc5"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.035561 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-inventory" (OuterVolumeSpecName: "inventory") pod "20c0c984-e49b-42f6-b3ee-385aa90befc5" (UID: "20c0c984-e49b-42f6-b3ee-385aa90befc5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.099913 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.100112 4944 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.100170 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.100257 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg5hb\" (UniqueName: \"kubernetes.io/projected/20c0c984-e49b-42f6-b3ee-385aa90befc5-kube-api-access-dg5hb\") on node \"crc\" DevicePath \"\"" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.100316 4944 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.100372 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/20c0c984-e49b-42f6-b3ee-385aa90befc5-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.201127 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" event={"ID":"20c0c984-e49b-42f6-b3ee-385aa90befc5","Type":"ContainerDied","Data":"67da0f7710e1a5f6ce30bfe252e3c9b438598fd7ee9044052d24ee2b96ef8aae"} Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.201170 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67da0f7710e1a5f6ce30bfe252e3c9b438598fd7ee9044052d24ee2b96ef8aae" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.201185 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-k5pgj" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.415212 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw"] Nov 24 11:15:10 crc kubenswrapper[4944]: E1124 11:15:10.415675 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a85ac49-8904-4650-961a-418e5144c293" containerName="collect-profiles" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.415687 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a85ac49-8904-4650-961a-418e5144c293" containerName="collect-profiles" Nov 24 11:15:10 crc kubenswrapper[4944]: E1124 11:15:10.415699 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c0c984-e49b-42f6-b3ee-385aa90befc5" containerName="neutron-sriov-openstack-openstack-cell1" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.415706 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c0c984-e49b-42f6-b3ee-385aa90befc5" containerName="neutron-sriov-openstack-openstack-cell1" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.415896 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="20c0c984-e49b-42f6-b3ee-385aa90befc5" containerName="neutron-sriov-openstack-openstack-cell1" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.415919 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a85ac49-8904-4650-961a-418e5144c293" containerName="collect-profiles" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.416660 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.419238 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.419710 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.419906 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.420938 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.421961 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.460114 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw"] Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.508329 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.508412 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.508466 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.508516 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.508540 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.508738 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mpzc\" (UniqueName: \"kubernetes.io/projected/3dea0418-b412-4bb6-856c-f34de24377a1-kube-api-access-2mpzc\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.611295 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mpzc\" (UniqueName: \"kubernetes.io/projected/3dea0418-b412-4bb6-856c-f34de24377a1-kube-api-access-2mpzc\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.611830 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.611860 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.611900 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.611918 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.611941 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.616674 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.616872 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.617568 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.617750 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.617977 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.635696 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mpzc\" (UniqueName: \"kubernetes.io/projected/3dea0418-b412-4bb6-856c-f34de24377a1-kube-api-access-2mpzc\") pod \"neutron-dhcp-openstack-openstack-cell1-mhrzw\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:10 crc kubenswrapper[4944]: I1124 11:15:10.767755 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:15:11 crc kubenswrapper[4944]: I1124 11:15:11.319320 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw"] Nov 24 11:15:12 crc kubenswrapper[4944]: I1124 11:15:12.221790 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" event={"ID":"3dea0418-b412-4bb6-856c-f34de24377a1","Type":"ContainerStarted","Data":"4153b969899f226d167c9209c88fed9571e7870cee3507d2e2ee107d6916d86b"} Nov 24 11:15:13 crc kubenswrapper[4944]: I1124 11:15:13.234010 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" event={"ID":"3dea0418-b412-4bb6-856c-f34de24377a1","Type":"ContainerStarted","Data":"f0b92f3c466707687ca6593c55702cd6c26a0a3cf85155798f9c6fa470dfc4be"} Nov 24 11:15:13 crc kubenswrapper[4944]: I1124 11:15:13.262363 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" podStartSLOduration=2.732297352 podStartE2EDuration="3.262345769s" podCreationTimestamp="2025-11-24 11:15:10 +0000 UTC" firstStartedPulling="2025-11-24 11:15:11.804627455 +0000 UTC m=+8572.339067957" lastFinishedPulling="2025-11-24 11:15:12.334675912 +0000 UTC m=+8572.869116374" observedRunningTime="2025-11-24 11:15:13.25850722 +0000 UTC m=+8573.792947692" watchObservedRunningTime="2025-11-24 11:15:13.262345769 +0000 UTC m=+8573.796786231" Nov 24 11:15:22 crc kubenswrapper[4944]: I1124 11:15:22.277501 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:15:22 crc kubenswrapper[4944]: E1124 11:15:22.278334 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:15:34 crc kubenswrapper[4944]: I1124 11:15:34.277489 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:15:34 crc kubenswrapper[4944]: E1124 11:15:34.278287 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:15:36 crc kubenswrapper[4944]: I1124 11:15:36.838879 4944 scope.go:117] "RemoveContainer" containerID="b4c0e8a0e7791002aa41ea42b6a8d15bbfac598725e2b5cb07cbeaacf90749f5" Nov 24 11:15:46 crc kubenswrapper[4944]: I1124 11:15:46.277949 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:15:46 crc kubenswrapper[4944]: E1124 11:15:46.278748 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:15:57 crc kubenswrapper[4944]: I1124 11:15:57.277511 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:15:57 crc kubenswrapper[4944]: E1124 11:15:57.278576 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:16:11 crc kubenswrapper[4944]: I1124 11:16:11.279370 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:16:11 crc kubenswrapper[4944]: E1124 11:16:11.281875 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:16:24 crc kubenswrapper[4944]: I1124 11:16:24.281512 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:16:24 crc kubenswrapper[4944]: E1124 11:16:24.282351 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:16:36 crc kubenswrapper[4944]: I1124 11:16:36.280387 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:16:36 crc kubenswrapper[4944]: E1124 11:16:36.281373 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:16:51 crc kubenswrapper[4944]: I1124 11:16:51.276974 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:16:51 crc kubenswrapper[4944]: E1124 11:16:51.277979 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:17:03 crc kubenswrapper[4944]: I1124 11:17:03.277995 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:17:03 crc kubenswrapper[4944]: E1124 11:17:03.279276 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:17:15 crc kubenswrapper[4944]: I1124 11:17:15.277082 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:17:15 crc kubenswrapper[4944]: E1124 11:17:15.277872 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:17:30 crc kubenswrapper[4944]: I1124 11:17:30.283425 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:17:30 crc kubenswrapper[4944]: E1124 11:17:30.284152 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:17:43 crc kubenswrapper[4944]: I1124 11:17:43.276638 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:17:43 crc kubenswrapper[4944]: E1124 11:17:43.277459 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:17:57 crc kubenswrapper[4944]: I1124 11:17:57.277461 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:17:57 crc kubenswrapper[4944]: E1124 11:17:57.278253 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:18:08 crc kubenswrapper[4944]: I1124 11:18:08.277947 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:18:08 crc kubenswrapper[4944]: E1124 11:18:08.279335 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:18:19 crc kubenswrapper[4944]: I1124 11:18:19.277350 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:18:19 crc kubenswrapper[4944]: E1124 11:18:19.278149 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:18:33 crc kubenswrapper[4944]: I1124 11:18:33.276810 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:18:34 crc kubenswrapper[4944]: I1124 11:18:34.299204 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"7d73d2457794efa51b32f46defdfb5e370f07c447756fe38b7313df14f567eea"} Nov 24 11:19:41 crc kubenswrapper[4944]: I1124 11:19:41.010165 4944 generic.go:334] "Generic (PLEG): container finished" podID="3dea0418-b412-4bb6-856c-f34de24377a1" containerID="f0b92f3c466707687ca6593c55702cd6c26a0a3cf85155798f9c6fa470dfc4be" exitCode=0 Nov 24 11:19:41 crc kubenswrapper[4944]: I1124 11:19:41.010238 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" event={"ID":"3dea0418-b412-4bb6-856c-f34de24377a1","Type":"ContainerDied","Data":"f0b92f3c466707687ca6593c55702cd6c26a0a3cf85155798f9c6fa470dfc4be"} Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.479127 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.604772 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-combined-ca-bundle\") pod \"3dea0418-b412-4bb6-856c-f34de24377a1\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.604834 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-agent-neutron-config-0\") pod \"3dea0418-b412-4bb6-856c-f34de24377a1\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.604948 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mpzc\" (UniqueName: \"kubernetes.io/projected/3dea0418-b412-4bb6-856c-f34de24377a1-kube-api-access-2mpzc\") pod \"3dea0418-b412-4bb6-856c-f34de24377a1\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.604977 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ceph\") pod \"3dea0418-b412-4bb6-856c-f34de24377a1\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.605024 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-inventory\") pod \"3dea0418-b412-4bb6-856c-f34de24377a1\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.605165 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ssh-key\") pod \"3dea0418-b412-4bb6-856c-f34de24377a1\" (UID: \"3dea0418-b412-4bb6-856c-f34de24377a1\") " Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.611379 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ceph" (OuterVolumeSpecName: "ceph") pod "3dea0418-b412-4bb6-856c-f34de24377a1" (UID: "3dea0418-b412-4bb6-856c-f34de24377a1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.614355 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "3dea0418-b412-4bb6-856c-f34de24377a1" (UID: "3dea0418-b412-4bb6-856c-f34de24377a1"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.616314 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dea0418-b412-4bb6-856c-f34de24377a1-kube-api-access-2mpzc" (OuterVolumeSpecName: "kube-api-access-2mpzc") pod "3dea0418-b412-4bb6-856c-f34de24377a1" (UID: "3dea0418-b412-4bb6-856c-f34de24377a1"). InnerVolumeSpecName "kube-api-access-2mpzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.648010 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "3dea0418-b412-4bb6-856c-f34de24377a1" (UID: "3dea0418-b412-4bb6-856c-f34de24377a1"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.649215 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3dea0418-b412-4bb6-856c-f34de24377a1" (UID: "3dea0418-b412-4bb6-856c-f34de24377a1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.652477 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-inventory" (OuterVolumeSpecName: "inventory") pod "3dea0418-b412-4bb6-856c-f34de24377a1" (UID: "3dea0418-b412-4bb6-856c-f34de24377a1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.708169 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.708458 4944 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.708470 4944 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.708483 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mpzc\" (UniqueName: \"kubernetes.io/projected/3dea0418-b412-4bb6-856c-f34de24377a1-kube-api-access-2mpzc\") on node \"crc\" DevicePath \"\"" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.708493 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 11:19:42 crc kubenswrapper[4944]: I1124 11:19:42.708501 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dea0418-b412-4bb6-856c-f34de24377a1-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 11:19:43 crc kubenswrapper[4944]: I1124 11:19:43.037160 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" event={"ID":"3dea0418-b412-4bb6-856c-f34de24377a1","Type":"ContainerDied","Data":"4153b969899f226d167c9209c88fed9571e7870cee3507d2e2ee107d6916d86b"} Nov 24 11:19:43 crc kubenswrapper[4944]: I1124 11:19:43.037231 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4153b969899f226d167c9209c88fed9571e7870cee3507d2e2ee107d6916d86b" Nov 24 11:19:43 crc kubenswrapper[4944]: I1124 11:19:43.037345 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-mhrzw" Nov 24 11:20:12 crc kubenswrapper[4944]: I1124 11:20:12.532419 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 11:20:12 crc kubenswrapper[4944]: I1124 11:20:12.533900 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="f7a89164-b59e-4690-b0e5-3482f54a769c" containerName="nova-cell0-conductor-conductor" containerID="cri-o://9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2" gracePeriod=30 Nov 24 11:20:13 crc kubenswrapper[4944]: I1124 11:20:13.093591 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 11:20:13 crc kubenswrapper[4944]: I1124 11:20:13.093950 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="e8551f1e-3692-4d2a-99b3-1d1015b35e9c" containerName="nova-cell1-conductor-conductor" containerID="cri-o://9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb" gracePeriod=30 Nov 24 11:20:13 crc kubenswrapper[4944]: I1124 11:20:13.237079 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 11:20:13 crc kubenswrapper[4944]: I1124 11:20:13.237702 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f028a568-aeb6-435b-81f5-d5014b3609c3" containerName="nova-scheduler-scheduler" containerID="cri-o://520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81" gracePeriod=30 Nov 24 11:20:13 crc kubenswrapper[4944]: I1124 11:20:13.246411 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 11:20:13 crc kubenswrapper[4944]: I1124 11:20:13.246723 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="64282688-4ab2-436d-bc65-97c100066cd2" containerName="nova-api-log" containerID="cri-o://651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676" gracePeriod=30 Nov 24 11:20:13 crc kubenswrapper[4944]: I1124 11:20:13.246820 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="64282688-4ab2-436d-bc65-97c100066cd2" containerName="nova-api-api" containerID="cri-o://1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8" gracePeriod=30 Nov 24 11:20:13 crc kubenswrapper[4944]: I1124 11:20:13.325881 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 11:20:13 crc kubenswrapper[4944]: I1124 11:20:13.326128 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-log" containerID="cri-o://1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6" gracePeriod=30 Nov 24 11:20:13 crc kubenswrapper[4944]: I1124 11:20:13.326662 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-metadata" containerID="cri-o://37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426" gracePeriod=30 Nov 24 11:20:14 crc kubenswrapper[4944]: I1124 11:20:14.381862 4944 generic.go:334] "Generic (PLEG): container finished" podID="64282688-4ab2-436d-bc65-97c100066cd2" containerID="651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676" exitCode=143 Nov 24 11:20:14 crc kubenswrapper[4944]: I1124 11:20:14.382172 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64282688-4ab2-436d-bc65-97c100066cd2","Type":"ContainerDied","Data":"651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676"} Nov 24 11:20:14 crc kubenswrapper[4944]: I1124 11:20:14.385062 4944 generic.go:334] "Generic (PLEG): container finished" podID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerID="1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6" exitCode=143 Nov 24 11:20:14 crc kubenswrapper[4944]: I1124 11:20:14.385084 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"880181b3-27ad-4f6c-9471-6016ea2d7e75","Type":"ContainerDied","Data":"1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6"} Nov 24 11:20:15 crc kubenswrapper[4944]: I1124 11:20:15.993252 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.001976 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.125173 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-combined-ca-bundle\") pod \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.125262 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-config-data\") pod \"f028a568-aeb6-435b-81f5-d5014b3609c3\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.125303 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-combined-ca-bundle\") pod \"f028a568-aeb6-435b-81f5-d5014b3609c3\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.125326 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw2hw\" (UniqueName: \"kubernetes.io/projected/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-kube-api-access-qw2hw\") pod \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.125380 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85r4p\" (UniqueName: \"kubernetes.io/projected/f028a568-aeb6-435b-81f5-d5014b3609c3-kube-api-access-85r4p\") pod \"f028a568-aeb6-435b-81f5-d5014b3609c3\" (UID: \"f028a568-aeb6-435b-81f5-d5014b3609c3\") " Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.125594 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-config-data\") pod \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\" (UID: \"e8551f1e-3692-4d2a-99b3-1d1015b35e9c\") " Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.130799 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-kube-api-access-qw2hw" (OuterVolumeSpecName: "kube-api-access-qw2hw") pod "e8551f1e-3692-4d2a-99b3-1d1015b35e9c" (UID: "e8551f1e-3692-4d2a-99b3-1d1015b35e9c"). InnerVolumeSpecName "kube-api-access-qw2hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.131113 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f028a568-aeb6-435b-81f5-d5014b3609c3-kube-api-access-85r4p" (OuterVolumeSpecName: "kube-api-access-85r4p") pod "f028a568-aeb6-435b-81f5-d5014b3609c3" (UID: "f028a568-aeb6-435b-81f5-d5014b3609c3"). InnerVolumeSpecName "kube-api-access-85r4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.153346 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f028a568-aeb6-435b-81f5-d5014b3609c3" (UID: "f028a568-aeb6-435b-81f5-d5014b3609c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.157728 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8551f1e-3692-4d2a-99b3-1d1015b35e9c" (UID: "e8551f1e-3692-4d2a-99b3-1d1015b35e9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.159272 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-config-data" (OuterVolumeSpecName: "config-data") pod "f028a568-aeb6-435b-81f5-d5014b3609c3" (UID: "f028a568-aeb6-435b-81f5-d5014b3609c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.163439 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-config-data" (OuterVolumeSpecName: "config-data") pod "e8551f1e-3692-4d2a-99b3-1d1015b35e9c" (UID: "e8551f1e-3692-4d2a-99b3-1d1015b35e9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.230272 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.230304 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.230314 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.230322 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f028a568-aeb6-435b-81f5-d5014b3609c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.230334 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw2hw\" (UniqueName: \"kubernetes.io/projected/e8551f1e-3692-4d2a-99b3-1d1015b35e9c-kube-api-access-qw2hw\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.230343 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85r4p\" (UniqueName: \"kubernetes.io/projected/f028a568-aeb6-435b-81f5-d5014b3609c3-kube-api-access-85r4p\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.391448 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.424334 4944 generic.go:334] "Generic (PLEG): container finished" podID="f7a89164-b59e-4690-b0e5-3482f54a769c" containerID="9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2" exitCode=0 Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.424388 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.424438 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f7a89164-b59e-4690-b0e5-3482f54a769c","Type":"ContainerDied","Data":"9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2"} Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.424495 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f7a89164-b59e-4690-b0e5-3482f54a769c","Type":"ContainerDied","Data":"1fe0ffbfd7d4e5cb3acedd720e2ea41d0dcac778901253bd7120db4841fd78ad"} Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.424514 4944 scope.go:117] "RemoveContainer" containerID="9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.439889 4944 generic.go:334] "Generic (PLEG): container finished" podID="f028a568-aeb6-435b-81f5-d5014b3609c3" containerID="520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81" exitCode=0 Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.439966 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f028a568-aeb6-435b-81f5-d5014b3609c3","Type":"ContainerDied","Data":"520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81"} Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.439985 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.439999 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f028a568-aeb6-435b-81f5-d5014b3609c3","Type":"ContainerDied","Data":"533851fee3ac440c0c311eb6b73c444e1655970aeb0d69e7e82c9dd7fcc498bd"} Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.445320 4944 generic.go:334] "Generic (PLEG): container finished" podID="e8551f1e-3692-4d2a-99b3-1d1015b35e9c" containerID="9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb" exitCode=0 Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.445414 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.445502 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e8551f1e-3692-4d2a-99b3-1d1015b35e9c","Type":"ContainerDied","Data":"9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb"} Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.445584 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e8551f1e-3692-4d2a-99b3-1d1015b35e9c","Type":"ContainerDied","Data":"c2238a3452b6725552bdcd3e6ab24ca43238d800d732c1e6c9c7c16a575109ea"} Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.460905 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": read tcp 10.217.0.2:43274->10.217.1.84:8775: read: connection reset by peer" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.461065 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": read tcp 10.217.0.2:43286->10.217.1.84:8775: read: connection reset by peer" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.481031 4944 scope.go:117] "RemoveContainer" containerID="9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2" Nov 24 11:20:16 crc kubenswrapper[4944]: E1124 11:20:16.482710 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2\": container with ID starting with 9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2 not found: ID does not exist" containerID="9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.482764 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2"} err="failed to get container status \"9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2\": rpc error: code = NotFound desc = could not find container \"9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2\": container with ID starting with 9027b00b1251679dd8a89db6d16e6965e3f6f748adb02fcea1d8f97ec72acea2 not found: ID does not exist" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.482799 4944 scope.go:117] "RemoveContainer" containerID="520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.493040 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.536846 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.538863 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-combined-ca-bundle\") pod \"f7a89164-b59e-4690-b0e5-3482f54a769c\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.538942 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-config-data\") pod \"f7a89164-b59e-4690-b0e5-3482f54a769c\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.539015 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7n99\" (UniqueName: \"kubernetes.io/projected/f7a89164-b59e-4690-b0e5-3482f54a769c-kube-api-access-c7n99\") pod \"f7a89164-b59e-4690-b0e5-3482f54a769c\" (UID: \"f7a89164-b59e-4690-b0e5-3482f54a769c\") " Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.543937 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7a89164-b59e-4690-b0e5-3482f54a769c-kube-api-access-c7n99" (OuterVolumeSpecName: "kube-api-access-c7n99") pod "f7a89164-b59e-4690-b0e5-3482f54a769c" (UID: "f7a89164-b59e-4690-b0e5-3482f54a769c"). InnerVolumeSpecName "kube-api-access-c7n99". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.545083 4944 scope.go:117] "RemoveContainer" containerID="520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.547035 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: E1124 11:20:16.547243 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81\": container with ID starting with 520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81 not found: ID does not exist" containerID="520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.547279 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81"} err="failed to get container status \"520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81\": rpc error: code = NotFound desc = could not find container \"520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81\": container with ID starting with 520cd08e52e1da77b78e5aebd222428a80dff92c8470ba5d9748730673da3a81 not found: ID does not exist" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.547305 4944 scope.go:117] "RemoveContainer" containerID="9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.557768 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.566723 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: E1124 11:20:16.567412 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8551f1e-3692-4d2a-99b3-1d1015b35e9c" containerName="nova-cell1-conductor-conductor" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.567435 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8551f1e-3692-4d2a-99b3-1d1015b35e9c" containerName="nova-cell1-conductor-conductor" Nov 24 11:20:16 crc kubenswrapper[4944]: E1124 11:20:16.567490 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f028a568-aeb6-435b-81f5-d5014b3609c3" containerName="nova-scheduler-scheduler" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.567499 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f028a568-aeb6-435b-81f5-d5014b3609c3" containerName="nova-scheduler-scheduler" Nov 24 11:20:16 crc kubenswrapper[4944]: E1124 11:20:16.567514 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dea0418-b412-4bb6-856c-f34de24377a1" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.567521 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dea0418-b412-4bb6-856c-f34de24377a1" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 24 11:20:16 crc kubenswrapper[4944]: E1124 11:20:16.567545 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7a89164-b59e-4690-b0e5-3482f54a769c" containerName="nova-cell0-conductor-conductor" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.567553 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7a89164-b59e-4690-b0e5-3482f54a769c" containerName="nova-cell0-conductor-conductor" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.567795 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8551f1e-3692-4d2a-99b3-1d1015b35e9c" containerName="nova-cell1-conductor-conductor" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.567813 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7a89164-b59e-4690-b0e5-3482f54a769c" containerName="nova-cell0-conductor-conductor" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.567837 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dea0418-b412-4bb6-856c-f34de24377a1" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.567860 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="f028a568-aeb6-435b-81f5-d5014b3609c3" containerName="nova-scheduler-scheduler" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.568819 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.571486 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.571719 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-config-data" (OuterVolumeSpecName: "config-data") pod "f7a89164-b59e-4690-b0e5-3482f54a769c" (UID: "f7a89164-b59e-4690-b0e5-3482f54a769c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.571845 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7a89164-b59e-4690-b0e5-3482f54a769c" (UID: "f7a89164-b59e-4690-b0e5-3482f54a769c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.579272 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.582541 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.584706 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.590981 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.605108 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.640958 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.640991 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a89164-b59e-4690-b0e5-3482f54a769c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.641001 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7n99\" (UniqueName: \"kubernetes.io/projected/f7a89164-b59e-4690-b0e5-3482f54a769c-kube-api-access-c7n99\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.712626 4944 scope.go:117] "RemoveContainer" containerID="9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb" Nov 24 11:20:16 crc kubenswrapper[4944]: E1124 11:20:16.715308 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb\": container with ID starting with 9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb not found: ID does not exist" containerID="9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.715354 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb"} err="failed to get container status \"9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb\": rpc error: code = NotFound desc = could not find container \"9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb\": container with ID starting with 9ec77137fa4e6a5102329539b28faf5d60ce22e5d65b70d4d5f5e68962bd92bb not found: ID does not exist" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.745118 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6564735-987c-4999-a5c2-d13e06e59f8a-config-data\") pod \"nova-scheduler-0\" (UID: \"d6564735-987c-4999-a5c2-d13e06e59f8a\") " pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.745273 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7c0120-2a92-4a79-b544-5022417126c3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bb7c0120-2a92-4a79-b544-5022417126c3\") " pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.746003 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7c0120-2a92-4a79-b544-5022417126c3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bb7c0120-2a92-4a79-b544-5022417126c3\") " pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.746144 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6564735-987c-4999-a5c2-d13e06e59f8a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6564735-987c-4999-a5c2-d13e06e59f8a\") " pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.746328 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzwgj\" (UniqueName: \"kubernetes.io/projected/d6564735-987c-4999-a5c2-d13e06e59f8a-kube-api-access-kzwgj\") pod \"nova-scheduler-0\" (UID: \"d6564735-987c-4999-a5c2-d13e06e59f8a\") " pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.746366 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhzvl\" (UniqueName: \"kubernetes.io/projected/bb7c0120-2a92-4a79-b544-5022417126c3-kube-api-access-fhzvl\") pod \"nova-cell1-conductor-0\" (UID: \"bb7c0120-2a92-4a79-b544-5022417126c3\") " pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.805003 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.831121 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.848217 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzwgj\" (UniqueName: \"kubernetes.io/projected/d6564735-987c-4999-a5c2-d13e06e59f8a-kube-api-access-kzwgj\") pod \"nova-scheduler-0\" (UID: \"d6564735-987c-4999-a5c2-d13e06e59f8a\") " pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.848254 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhzvl\" (UniqueName: \"kubernetes.io/projected/bb7c0120-2a92-4a79-b544-5022417126c3-kube-api-access-fhzvl\") pod \"nova-cell1-conductor-0\" (UID: \"bb7c0120-2a92-4a79-b544-5022417126c3\") " pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.848330 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6564735-987c-4999-a5c2-d13e06e59f8a-config-data\") pod \"nova-scheduler-0\" (UID: \"d6564735-987c-4999-a5c2-d13e06e59f8a\") " pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.848396 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7c0120-2a92-4a79-b544-5022417126c3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bb7c0120-2a92-4a79-b544-5022417126c3\") " pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.848432 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7c0120-2a92-4a79-b544-5022417126c3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bb7c0120-2a92-4a79-b544-5022417126c3\") " pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.848455 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6564735-987c-4999-a5c2-d13e06e59f8a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6564735-987c-4999-a5c2-d13e06e59f8a\") " pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.855110 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.856662 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.860491 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.862461 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.868252 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6564735-987c-4999-a5c2-d13e06e59f8a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6564735-987c-4999-a5c2-d13e06e59f8a\") " pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.870926 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7c0120-2a92-4a79-b544-5022417126c3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bb7c0120-2a92-4a79-b544-5022417126c3\") " pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.872972 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb7c0120-2a92-4a79-b544-5022417126c3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bb7c0120-2a92-4a79-b544-5022417126c3\") " pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.873411 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhzvl\" (UniqueName: \"kubernetes.io/projected/bb7c0120-2a92-4a79-b544-5022417126c3-kube-api-access-fhzvl\") pod \"nova-cell1-conductor-0\" (UID: \"bb7c0120-2a92-4a79-b544-5022417126c3\") " pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.874096 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6564735-987c-4999-a5c2-d13e06e59f8a-config-data\") pod \"nova-scheduler-0\" (UID: \"d6564735-987c-4999-a5c2-d13e06e59f8a\") " pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.875160 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzwgj\" (UniqueName: \"kubernetes.io/projected/d6564735-987c-4999-a5c2-d13e06e59f8a-kube-api-access-kzwgj\") pod \"nova-scheduler-0\" (UID: \"d6564735-987c-4999-a5c2-d13e06e59f8a\") " pod="openstack/nova-scheduler-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.951524 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2142325c-7d11-49dd-ba85-e38b4da99c40-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2142325c-7d11-49dd-ba85-e38b4da99c40\") " pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.951615 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2142325c-7d11-49dd-ba85-e38b4da99c40-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2142325c-7d11-49dd-ba85-e38b4da99c40\") " pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.951812 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xptgd\" (UniqueName: \"kubernetes.io/projected/2142325c-7d11-49dd-ba85-e38b4da99c40-kube-api-access-xptgd\") pod \"nova-cell0-conductor-0\" (UID: \"2142325c-7d11-49dd-ba85-e38b4da99c40\") " pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:16 crc kubenswrapper[4944]: I1124 11:20:16.989628 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.000320 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.053319 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xptgd\" (UniqueName: \"kubernetes.io/projected/2142325c-7d11-49dd-ba85-e38b4da99c40-kube-api-access-xptgd\") pod \"nova-cell0-conductor-0\" (UID: \"2142325c-7d11-49dd-ba85-e38b4da99c40\") " pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.056104 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2142325c-7d11-49dd-ba85-e38b4da99c40-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2142325c-7d11-49dd-ba85-e38b4da99c40\") " pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.056203 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2142325c-7d11-49dd-ba85-e38b4da99c40-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2142325c-7d11-49dd-ba85-e38b4da99c40\") " pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.060833 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2142325c-7d11-49dd-ba85-e38b4da99c40-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2142325c-7d11-49dd-ba85-e38b4da99c40\") " pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.060918 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2142325c-7d11-49dd-ba85-e38b4da99c40-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2142325c-7d11-49dd-ba85-e38b4da99c40\") " pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.067454 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xptgd\" (UniqueName: \"kubernetes.io/projected/2142325c-7d11-49dd-ba85-e38b4da99c40-kube-api-access-xptgd\") pod \"nova-cell0-conductor-0\" (UID: \"2142325c-7d11-49dd-ba85-e38b4da99c40\") " pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.067553 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.074827 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.261470 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-combined-ca-bundle\") pod \"880181b3-27ad-4f6c-9471-6016ea2d7e75\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.261861 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnsnn\" (UniqueName: \"kubernetes.io/projected/64282688-4ab2-436d-bc65-97c100066cd2-kube-api-access-fnsnn\") pod \"64282688-4ab2-436d-bc65-97c100066cd2\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.261924 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-config-data\") pod \"64282688-4ab2-436d-bc65-97c100066cd2\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.261985 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-combined-ca-bundle\") pod \"64282688-4ab2-436d-bc65-97c100066cd2\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.262068 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-config-data\") pod \"880181b3-27ad-4f6c-9471-6016ea2d7e75\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.262144 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/880181b3-27ad-4f6c-9471-6016ea2d7e75-logs\") pod \"880181b3-27ad-4f6c-9471-6016ea2d7e75\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.262179 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g59cv\" (UniqueName: \"kubernetes.io/projected/880181b3-27ad-4f6c-9471-6016ea2d7e75-kube-api-access-g59cv\") pod \"880181b3-27ad-4f6c-9471-6016ea2d7e75\" (UID: \"880181b3-27ad-4f6c-9471-6016ea2d7e75\") " Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.262302 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64282688-4ab2-436d-bc65-97c100066cd2-logs\") pod \"64282688-4ab2-436d-bc65-97c100066cd2\" (UID: \"64282688-4ab2-436d-bc65-97c100066cd2\") " Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.263367 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64282688-4ab2-436d-bc65-97c100066cd2-logs" (OuterVolumeSpecName: "logs") pod "64282688-4ab2-436d-bc65-97c100066cd2" (UID: "64282688-4ab2-436d-bc65-97c100066cd2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.265091 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/880181b3-27ad-4f6c-9471-6016ea2d7e75-logs" (OuterVolumeSpecName: "logs") pod "880181b3-27ad-4f6c-9471-6016ea2d7e75" (UID: "880181b3-27ad-4f6c-9471-6016ea2d7e75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.282823 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/880181b3-27ad-4f6c-9471-6016ea2d7e75-kube-api-access-g59cv" (OuterVolumeSpecName: "kube-api-access-g59cv") pod "880181b3-27ad-4f6c-9471-6016ea2d7e75" (UID: "880181b3-27ad-4f6c-9471-6016ea2d7e75"). InnerVolumeSpecName "kube-api-access-g59cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.287127 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64282688-4ab2-436d-bc65-97c100066cd2-kube-api-access-fnsnn" (OuterVolumeSpecName: "kube-api-access-fnsnn") pod "64282688-4ab2-436d-bc65-97c100066cd2" (UID: "64282688-4ab2-436d-bc65-97c100066cd2"). InnerVolumeSpecName "kube-api-access-fnsnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.295275 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-config-data" (OuterVolumeSpecName: "config-data") pod "880181b3-27ad-4f6c-9471-6016ea2d7e75" (UID: "880181b3-27ad-4f6c-9471-6016ea2d7e75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.297868 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-config-data" (OuterVolumeSpecName: "config-data") pod "64282688-4ab2-436d-bc65-97c100066cd2" (UID: "64282688-4ab2-436d-bc65-97c100066cd2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.303248 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64282688-4ab2-436d-bc65-97c100066cd2" (UID: "64282688-4ab2-436d-bc65-97c100066cd2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.322688 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "880181b3-27ad-4f6c-9471-6016ea2d7e75" (UID: "880181b3-27ad-4f6c-9471-6016ea2d7e75"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.362592 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.364422 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.364448 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/880181b3-27ad-4f6c-9471-6016ea2d7e75-logs\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.364461 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g59cv\" (UniqueName: \"kubernetes.io/projected/880181b3-27ad-4f6c-9471-6016ea2d7e75-kube-api-access-g59cv\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.364471 4944 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64282688-4ab2-436d-bc65-97c100066cd2-logs\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.364479 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880181b3-27ad-4f6c-9471-6016ea2d7e75-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.364489 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnsnn\" (UniqueName: \"kubernetes.io/projected/64282688-4ab2-436d-bc65-97c100066cd2-kube-api-access-fnsnn\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.364497 4944 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.364505 4944 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64282688-4ab2-436d-bc65-97c100066cd2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.469071 4944 generic.go:334] "Generic (PLEG): container finished" podID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerID="37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426" exitCode=0 Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.469142 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"880181b3-27ad-4f6c-9471-6016ea2d7e75","Type":"ContainerDied","Data":"37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426"} Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.469174 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"880181b3-27ad-4f6c-9471-6016ea2d7e75","Type":"ContainerDied","Data":"ddbf2b9d8f7861ead89a5dc5d30a4a698e01b118d6ea1855b27ded05026526bb"} Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.469193 4944 scope.go:117] "RemoveContainer" containerID="37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.469329 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.476788 4944 generic.go:334] "Generic (PLEG): container finished" podID="64282688-4ab2-436d-bc65-97c100066cd2" containerID="1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8" exitCode=0 Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.476854 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64282688-4ab2-436d-bc65-97c100066cd2","Type":"ContainerDied","Data":"1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8"} Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.476878 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64282688-4ab2-436d-bc65-97c100066cd2","Type":"ContainerDied","Data":"9c9dca476d4e35e27c127d04c21a6059576314298d613acf33f14688da7c5550"} Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.476888 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.492312 4944 scope.go:117] "RemoveContainer" containerID="1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.521947 4944 scope.go:117] "RemoveContainer" containerID="37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426" Nov 24 11:20:17 crc kubenswrapper[4944]: E1124 11:20:17.522492 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426\": container with ID starting with 37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426 not found: ID does not exist" containerID="37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.522543 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426"} err="failed to get container status \"37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426\": rpc error: code = NotFound desc = could not find container \"37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426\": container with ID starting with 37ca9db9c7430fadc77afda1fab0a4bb45e78371c001ad006dbddb14ca39a426 not found: ID does not exist" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.522577 4944 scope.go:117] "RemoveContainer" containerID="1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6" Nov 24 11:20:17 crc kubenswrapper[4944]: E1124 11:20:17.522898 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6\": container with ID starting with 1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6 not found: ID does not exist" containerID="1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.522924 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6"} err="failed to get container status \"1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6\": rpc error: code = NotFound desc = could not find container \"1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6\": container with ID starting with 1d9c82f3f832400a5ef23a0f6742e810d6104f763d8e3bec8406fdf6dc9d22d6 not found: ID does not exist" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.522945 4944 scope.go:117] "RemoveContainer" containerID="1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.526937 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.538064 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.552508 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.562995 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.569744 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 11:20:17 crc kubenswrapper[4944]: E1124 11:20:17.570357 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64282688-4ab2-436d-bc65-97c100066cd2" containerName="nova-api-api" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.570375 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="64282688-4ab2-436d-bc65-97c100066cd2" containerName="nova-api-api" Nov 24 11:20:17 crc kubenswrapper[4944]: E1124 11:20:17.570416 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64282688-4ab2-436d-bc65-97c100066cd2" containerName="nova-api-log" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.570426 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="64282688-4ab2-436d-bc65-97c100066cd2" containerName="nova-api-log" Nov 24 11:20:17 crc kubenswrapper[4944]: E1124 11:20:17.570435 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-metadata" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.570443 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-metadata" Nov 24 11:20:17 crc kubenswrapper[4944]: E1124 11:20:17.570462 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-log" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.570469 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-log" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.570699 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-log" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.570716 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="64282688-4ab2-436d-bc65-97c100066cd2" containerName="nova-api-log" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.570735 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="64282688-4ab2-436d-bc65-97c100066cd2" containerName="nova-api-api" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.570755 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" containerName="nova-metadata-metadata" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.573707 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.577935 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.581221 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.583213 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.585110 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.592367 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.595549 4944 scope.go:117] "RemoveContainer" containerID="651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.602879 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 11:20:17 crc kubenswrapper[4944]: W1124 11:20:17.622509 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb7c0120_2a92_4a79_b544_5022417126c3.slice/crio-cb54e9ab8ffbc682e867338e9e2f1855ff277345e69e057bf22b5bf9297c148b WatchSource:0}: Error finding container cb54e9ab8ffbc682e867338e9e2f1855ff277345e69e057bf22b5bf9297c148b: Status 404 returned error can't find the container with id cb54e9ab8ffbc682e867338e9e2f1855ff277345e69e057bf22b5bf9297c148b Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.626316 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.657289 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 11:20:17 crc kubenswrapper[4944]: W1124 11:20:17.672700 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6564735_987c_4999_a5c2_d13e06e59f8a.slice/crio-ad0f6e996289e3abe0a69900f1ff800defa70f12a429b7dfca7d720d418850f5 WatchSource:0}: Error finding container ad0f6e996289e3abe0a69900f1ff800defa70f12a429b7dfca7d720d418850f5: Status 404 returned error can't find the container with id ad0f6e996289e3abe0a69900f1ff800defa70f12a429b7dfca7d720d418850f5 Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.674158 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a521fc51-f8b9-426d-a066-735ce6868d6f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.674264 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8da4376-b2e8-477b-b088-110d4c143a06-logs\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.674423 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a521fc51-f8b9-426d-a066-735ce6868d6f-logs\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.674471 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a521fc51-f8b9-426d-a066-735ce6868d6f-config-data\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.674569 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8da4376-b2e8-477b-b088-110d4c143a06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.674629 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt7h4\" (UniqueName: \"kubernetes.io/projected/a521fc51-f8b9-426d-a066-735ce6868d6f-kube-api-access-kt7h4\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.674839 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8da4376-b2e8-477b-b088-110d4c143a06-config-data\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.675067 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75p9v\" (UniqueName: \"kubernetes.io/projected/a8da4376-b2e8-477b-b088-110d4c143a06-kube-api-access-75p9v\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.777288 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75p9v\" (UniqueName: \"kubernetes.io/projected/a8da4376-b2e8-477b-b088-110d4c143a06-kube-api-access-75p9v\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.777344 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a521fc51-f8b9-426d-a066-735ce6868d6f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.777411 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8da4376-b2e8-477b-b088-110d4c143a06-logs\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.777468 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a521fc51-f8b9-426d-a066-735ce6868d6f-config-data\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.777483 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a521fc51-f8b9-426d-a066-735ce6868d6f-logs\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.777550 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8da4376-b2e8-477b-b088-110d4c143a06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.777590 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt7h4\" (UniqueName: \"kubernetes.io/projected/a521fc51-f8b9-426d-a066-735ce6868d6f-kube-api-access-kt7h4\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.777823 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8da4376-b2e8-477b-b088-110d4c143a06-config-data\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.779208 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a521fc51-f8b9-426d-a066-735ce6868d6f-logs\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.779608 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8da4376-b2e8-477b-b088-110d4c143a06-logs\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.782682 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a521fc51-f8b9-426d-a066-735ce6868d6f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.783270 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a521fc51-f8b9-426d-a066-735ce6868d6f-config-data\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.783789 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8da4376-b2e8-477b-b088-110d4c143a06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.785095 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8da4376-b2e8-477b-b088-110d4c143a06-config-data\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.796576 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75p9v\" (UniqueName: \"kubernetes.io/projected/a8da4376-b2e8-477b-b088-110d4c143a06-kube-api-access-75p9v\") pod \"nova-metadata-0\" (UID: \"a8da4376-b2e8-477b-b088-110d4c143a06\") " pod="openstack/nova-metadata-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.799152 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt7h4\" (UniqueName: \"kubernetes.io/projected/a521fc51-f8b9-426d-a066-735ce6868d6f-kube-api-access-kt7h4\") pod \"nova-api-0\" (UID: \"a521fc51-f8b9-426d-a066-735ce6868d6f\") " pod="openstack/nova-api-0" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.876609 4944 scope.go:117] "RemoveContainer" containerID="1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8" Nov 24 11:20:17 crc kubenswrapper[4944]: E1124 11:20:17.877125 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8\": container with ID starting with 1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8 not found: ID does not exist" containerID="1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.877161 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8"} err="failed to get container status \"1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8\": rpc error: code = NotFound desc = could not find container \"1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8\": container with ID starting with 1444b8bd654ff4458bf47b796612318df495bd0456bcf3fe3ef5195c10f0a9b8 not found: ID does not exist" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.877184 4944 scope.go:117] "RemoveContainer" containerID="651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676" Nov 24 11:20:17 crc kubenswrapper[4944]: E1124 11:20:17.877498 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676\": container with ID starting with 651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676 not found: ID does not exist" containerID="651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.877664 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676"} err="failed to get container status \"651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676\": rpc error: code = NotFound desc = could not find container \"651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676\": container with ID starting with 651e38b23e4c4b5a986d582cd6c48c4f6a8f5623065a6f9aafdf2d9f5012a676 not found: ID does not exist" Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.891351 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 11:20:17 crc kubenswrapper[4944]: I1124 11:20:17.991292 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.032377 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.326068 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64282688-4ab2-436d-bc65-97c100066cd2" path="/var/lib/kubelet/pods/64282688-4ab2-436d-bc65-97c100066cd2/volumes" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.328563 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="880181b3-27ad-4f6c-9471-6016ea2d7e75" path="/var/lib/kubelet/pods/880181b3-27ad-4f6c-9471-6016ea2d7e75/volumes" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.329158 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8551f1e-3692-4d2a-99b3-1d1015b35e9c" path="/var/lib/kubelet/pods/e8551f1e-3692-4d2a-99b3-1d1015b35e9c/volumes" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.331517 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f028a568-aeb6-435b-81f5-d5014b3609c3" path="/var/lib/kubelet/pods/f028a568-aeb6-435b-81f5-d5014b3609c3/volumes" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.332043 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7a89164-b59e-4690-b0e5-3482f54a769c" path="/var/lib/kubelet/pods/f7a89164-b59e-4690-b0e5-3482f54a769c/volumes" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.482251 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 11:20:18 crc kubenswrapper[4944]: W1124 11:20:18.486385 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8da4376_b2e8_477b_b088_110d4c143a06.slice/crio-fe78950e9c774a2d51f0aafea00c5f1431bdc102f8121bd8db688e6236e8890d WatchSource:0}: Error finding container fe78950e9c774a2d51f0aafea00c5f1431bdc102f8121bd8db688e6236e8890d: Status 404 returned error can't find the container with id fe78950e9c774a2d51f0aafea00c5f1431bdc102f8121bd8db688e6236e8890d Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.486611 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bb7c0120-2a92-4a79-b544-5022417126c3","Type":"ContainerStarted","Data":"58d18fd70d42a16b16cb541e4e3d03a715de51151e1ed5ae010e7aafd646af22"} Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.486660 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bb7c0120-2a92-4a79-b544-5022417126c3","Type":"ContainerStarted","Data":"cb54e9ab8ffbc682e867338e9e2f1855ff277345e69e057bf22b5bf9297c148b"} Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.487070 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.492532 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2142325c-7d11-49dd-ba85-e38b4da99c40","Type":"ContainerStarted","Data":"4d2f3471cd8d489d683bb6896473e6b44f526aef3104eb7f0cb84b049039de0f"} Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.492575 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2142325c-7d11-49dd-ba85-e38b4da99c40","Type":"ContainerStarted","Data":"51e53686edf707f019d9453ccd24672ca0234a1c8f7ba15e73186cbbb9f316c5"} Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.492628 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.495628 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d6564735-987c-4999-a5c2-d13e06e59f8a","Type":"ContainerStarted","Data":"e3b963bd0bd7c336660f49459cd9486e984dce184629adb983801e7b3b39d8ff"} Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.495726 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d6564735-987c-4999-a5c2-d13e06e59f8a","Type":"ContainerStarted","Data":"ad0f6e996289e3abe0a69900f1ff800defa70f12a429b7dfca7d720d418850f5"} Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.531765 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.531740591 podStartE2EDuration="2.531740591s" podCreationTimestamp="2025-11-24 11:20:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 11:20:18.510020956 +0000 UTC m=+8879.044461418" watchObservedRunningTime="2025-11-24 11:20:18.531740591 +0000 UTC m=+8879.066181053" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.535923 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.535905982 podStartE2EDuration="2.535905982s" podCreationTimestamp="2025-11-24 11:20:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 11:20:18.53015409 +0000 UTC m=+8879.064594572" watchObservedRunningTime="2025-11-24 11:20:18.535905982 +0000 UTC m=+8879.070346444" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.551245 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.551222665 podStartE2EDuration="2.551222665s" podCreationTimestamp="2025-11-24 11:20:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 11:20:18.546543278 +0000 UTC m=+8879.080983740" watchObservedRunningTime="2025-11-24 11:20:18.551222665 +0000 UTC m=+8879.085663127" Nov 24 11:20:18 crc kubenswrapper[4944]: I1124 11:20:18.631268 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 11:20:19 crc kubenswrapper[4944]: I1124 11:20:19.509873 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a8da4376-b2e8-477b-b088-110d4c143a06","Type":"ContainerStarted","Data":"7057052e325d052035f7ac8b2fc209804339e5cb3fb26039cbec23e96392a126"} Nov 24 11:20:19 crc kubenswrapper[4944]: I1124 11:20:19.510340 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a8da4376-b2e8-477b-b088-110d4c143a06","Type":"ContainerStarted","Data":"fe78950e9c774a2d51f0aafea00c5f1431bdc102f8121bd8db688e6236e8890d"} Nov 24 11:20:19 crc kubenswrapper[4944]: I1124 11:20:19.514202 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a521fc51-f8b9-426d-a066-735ce6868d6f","Type":"ContainerStarted","Data":"6dbb80688cf579a0abf57a356eed74d5a36a7060e39d70be5c881896f4be01d9"} Nov 24 11:20:19 crc kubenswrapper[4944]: I1124 11:20:19.514271 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a521fc51-f8b9-426d-a066-735ce6868d6f","Type":"ContainerStarted","Data":"46afcdae0c6931a38dae08dc2546f074c2716fdd6de0b0dff8778242b75740a4"} Nov 24 11:20:20 crc kubenswrapper[4944]: I1124 11:20:20.529201 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a8da4376-b2e8-477b-b088-110d4c143a06","Type":"ContainerStarted","Data":"65a96963cf8be0c9abd8218db40f6edbb68e65e38a691ab4a50432087f0140fd"} Nov 24 11:20:20 crc kubenswrapper[4944]: I1124 11:20:20.531505 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a521fc51-f8b9-426d-a066-735ce6868d6f","Type":"ContainerStarted","Data":"dee6387d128f1336d16804d6632a2a6c083bb6b70e89be21fc1bfafec1f93838"} Nov 24 11:20:20 crc kubenswrapper[4944]: I1124 11:20:20.581209 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.581188269 podStartE2EDuration="3.581188269s" podCreationTimestamp="2025-11-24 11:20:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 11:20:20.571764941 +0000 UTC m=+8881.106205413" watchObservedRunningTime="2025-11-24 11:20:20.581188269 +0000 UTC m=+8881.115628731" Nov 24 11:20:20 crc kubenswrapper[4944]: I1124 11:20:20.581479 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.581471698 podStartE2EDuration="3.581471698s" podCreationTimestamp="2025-11-24 11:20:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 11:20:20.555111506 +0000 UTC m=+8881.089551988" watchObservedRunningTime="2025-11-24 11:20:20.581471698 +0000 UTC m=+8881.115912180" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.375856 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dzn89"] Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.380018 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.387020 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzn89"] Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.488281 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-catalog-content\") pod \"redhat-marketplace-dzn89\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.488682 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-utilities\") pod \"redhat-marketplace-dzn89\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.488921 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjwpt\" (UniqueName: \"kubernetes.io/projected/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-kube-api-access-zjwpt\") pod \"redhat-marketplace-dzn89\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.591499 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-catalog-content\") pod \"redhat-marketplace-dzn89\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.591637 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-utilities\") pod \"redhat-marketplace-dzn89\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.591711 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjwpt\" (UniqueName: \"kubernetes.io/projected/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-kube-api-access-zjwpt\") pod \"redhat-marketplace-dzn89\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.592267 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-catalog-content\") pod \"redhat-marketplace-dzn89\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.592368 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-utilities\") pod \"redhat-marketplace-dzn89\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.630559 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjwpt\" (UniqueName: \"kubernetes.io/projected/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-kube-api-access-zjwpt\") pod \"redhat-marketplace-dzn89\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.709781 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:21 crc kubenswrapper[4944]: I1124 11:20:21.990895 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 11:20:22 crc kubenswrapper[4944]: I1124 11:20:22.189393 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzn89"] Nov 24 11:20:22 crc kubenswrapper[4944]: W1124 11:20:22.191185 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ebf2b3e_cbb0_4b9b_bb3a_6d2d089b38b5.slice/crio-2bb3f36c011cb2a19157073f98e7dce4816a282db66b1e20d717593167d99ecb WatchSource:0}: Error finding container 2bb3f36c011cb2a19157073f98e7dce4816a282db66b1e20d717593167d99ecb: Status 404 returned error can't find the container with id 2bb3f36c011cb2a19157073f98e7dce4816a282db66b1e20d717593167d99ecb Nov 24 11:20:22 crc kubenswrapper[4944]: I1124 11:20:22.554764 4944 generic.go:334] "Generic (PLEG): container finished" podID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" containerID="e37172644dfc889d3ffdf042fa0691f5b4715c8f9da4e2d6af8400e94e5fd96a" exitCode=0 Nov 24 11:20:22 crc kubenswrapper[4944]: I1124 11:20:22.554826 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzn89" event={"ID":"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5","Type":"ContainerDied","Data":"e37172644dfc889d3ffdf042fa0691f5b4715c8f9da4e2d6af8400e94e5fd96a"} Nov 24 11:20:22 crc kubenswrapper[4944]: I1124 11:20:22.555156 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzn89" event={"ID":"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5","Type":"ContainerStarted","Data":"2bb3f36c011cb2a19157073f98e7dce4816a282db66b1e20d717593167d99ecb"} Nov 24 11:20:22 crc kubenswrapper[4944]: I1124 11:20:22.557954 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 11:20:22 crc kubenswrapper[4944]: I1124 11:20:22.992093 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 11:20:22 crc kubenswrapper[4944]: I1124 11:20:22.993429 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 11:20:23 crc kubenswrapper[4944]: I1124 11:20:23.567865 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzn89" event={"ID":"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5","Type":"ContainerStarted","Data":"adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09"} Nov 24 11:20:24 crc kubenswrapper[4944]: I1124 11:20:24.579921 4944 generic.go:334] "Generic (PLEG): container finished" podID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" containerID="adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09" exitCode=0 Nov 24 11:20:24 crc kubenswrapper[4944]: I1124 11:20:24.580058 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzn89" event={"ID":"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5","Type":"ContainerDied","Data":"adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09"} Nov 24 11:20:25 crc kubenswrapper[4944]: I1124 11:20:25.592558 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzn89" event={"ID":"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5","Type":"ContainerStarted","Data":"d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c"} Nov 24 11:20:25 crc kubenswrapper[4944]: I1124 11:20:25.619418 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dzn89" podStartSLOduration=2.204529583 podStartE2EDuration="4.619396319s" podCreationTimestamp="2025-11-24 11:20:21 +0000 UTC" firstStartedPulling="2025-11-24 11:20:22.557545571 +0000 UTC m=+8883.091986033" lastFinishedPulling="2025-11-24 11:20:24.972412277 +0000 UTC m=+8885.506852769" observedRunningTime="2025-11-24 11:20:25.615582859 +0000 UTC m=+8886.150023311" watchObservedRunningTime="2025-11-24 11:20:25.619396319 +0000 UTC m=+8886.153836791" Nov 24 11:20:26 crc kubenswrapper[4944]: I1124 11:20:26.990513 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 11:20:27 crc kubenswrapper[4944]: I1124 11:20:27.019601 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 11:20:27 crc kubenswrapper[4944]: I1124 11:20:27.037233 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 11:20:27 crc kubenswrapper[4944]: I1124 11:20:27.400283 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 11:20:27 crc kubenswrapper[4944]: I1124 11:20:27.643124 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 11:20:27 crc kubenswrapper[4944]: I1124 11:20:27.991708 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 11:20:27 crc kubenswrapper[4944]: I1124 11:20:27.992362 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 11:20:28 crc kubenswrapper[4944]: I1124 11:20:28.033711 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 11:20:28 crc kubenswrapper[4944]: I1124 11:20:28.034331 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 11:20:29 crc kubenswrapper[4944]: I1124 11:20:29.074365 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a8da4376-b2e8-477b-b088-110d4c143a06" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.191:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 11:20:29 crc kubenswrapper[4944]: I1124 11:20:29.074352 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a8da4376-b2e8-477b-b088-110d4c143a06" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.191:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 11:20:29 crc kubenswrapper[4944]: I1124 11:20:29.161013 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a521fc51-f8b9-426d-a066-735ce6868d6f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 11:20:29 crc kubenswrapper[4944]: I1124 11:20:29.161034 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a521fc51-f8b9-426d-a066-735ce6868d6f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 11:20:31 crc kubenswrapper[4944]: I1124 11:20:31.710552 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:31 crc kubenswrapper[4944]: I1124 11:20:31.710955 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:31 crc kubenswrapper[4944]: I1124 11:20:31.785965 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:33 crc kubenswrapper[4944]: I1124 11:20:33.041019 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:33 crc kubenswrapper[4944]: I1124 11:20:33.106252 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzn89"] Nov 24 11:20:34 crc kubenswrapper[4944]: I1124 11:20:34.726678 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dzn89" podUID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" containerName="registry-server" containerID="cri-o://d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c" gracePeriod=2 Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.272181 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.365397 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-catalog-content\") pod \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.365601 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjwpt\" (UniqueName: \"kubernetes.io/projected/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-kube-api-access-zjwpt\") pod \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.365651 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-utilities\") pod \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\" (UID: \"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5\") " Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.368066 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-utilities" (OuterVolumeSpecName: "utilities") pod "6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" (UID: "6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.373770 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-kube-api-access-zjwpt" (OuterVolumeSpecName: "kube-api-access-zjwpt") pod "6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" (UID: "6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5"). InnerVolumeSpecName "kube-api-access-zjwpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.384592 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" (UID: "6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.473530 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjwpt\" (UniqueName: \"kubernetes.io/projected/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-kube-api-access-zjwpt\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.473582 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.473596 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.738070 4944 generic.go:334] "Generic (PLEG): container finished" podID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" containerID="d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c" exitCode=0 Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.738110 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzn89" event={"ID":"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5","Type":"ContainerDied","Data":"d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c"} Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.738140 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzn89" event={"ID":"6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5","Type":"ContainerDied","Data":"2bb3f36c011cb2a19157073f98e7dce4816a282db66b1e20d717593167d99ecb"} Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.738157 4944 scope.go:117] "RemoveContainer" containerID="d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.738153 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzn89" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.785968 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzn89"] Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.790399 4944 scope.go:117] "RemoveContainer" containerID="adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.800170 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzn89"] Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.825484 4944 scope.go:117] "RemoveContainer" containerID="e37172644dfc889d3ffdf042fa0691f5b4715c8f9da4e2d6af8400e94e5fd96a" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.876145 4944 scope.go:117] "RemoveContainer" containerID="d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c" Nov 24 11:20:35 crc kubenswrapper[4944]: E1124 11:20:35.876923 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c\": container with ID starting with d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c not found: ID does not exist" containerID="d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.876975 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c"} err="failed to get container status \"d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c\": rpc error: code = NotFound desc = could not find container \"d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c\": container with ID starting with d43dde1effef330c31518da6dfa71b65d0bba2384ee727f255f96bbaff2e303c not found: ID does not exist" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.877011 4944 scope.go:117] "RemoveContainer" containerID="adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09" Nov 24 11:20:35 crc kubenswrapper[4944]: E1124 11:20:35.877660 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09\": container with ID starting with adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09 not found: ID does not exist" containerID="adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.877712 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09"} err="failed to get container status \"adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09\": rpc error: code = NotFound desc = could not find container \"adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09\": container with ID starting with adc2893c9db60d8c430169a19642cbb643716dbe520732aa3c56a654ecc37c09 not found: ID does not exist" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.877746 4944 scope.go:117] "RemoveContainer" containerID="e37172644dfc889d3ffdf042fa0691f5b4715c8f9da4e2d6af8400e94e5fd96a" Nov 24 11:20:35 crc kubenswrapper[4944]: E1124 11:20:35.878181 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e37172644dfc889d3ffdf042fa0691f5b4715c8f9da4e2d6af8400e94e5fd96a\": container with ID starting with e37172644dfc889d3ffdf042fa0691f5b4715c8f9da4e2d6af8400e94e5fd96a not found: ID does not exist" containerID="e37172644dfc889d3ffdf042fa0691f5b4715c8f9da4e2d6af8400e94e5fd96a" Nov 24 11:20:35 crc kubenswrapper[4944]: I1124 11:20:35.878245 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e37172644dfc889d3ffdf042fa0691f5b4715c8f9da4e2d6af8400e94e5fd96a"} err="failed to get container status \"e37172644dfc889d3ffdf042fa0691f5b4715c8f9da4e2d6af8400e94e5fd96a\": rpc error: code = NotFound desc = could not find container \"e37172644dfc889d3ffdf042fa0691f5b4715c8f9da4e2d6af8400e94e5fd96a\": container with ID starting with e37172644dfc889d3ffdf042fa0691f5b4715c8f9da4e2d6af8400e94e5fd96a not found: ID does not exist" Nov 24 11:20:36 crc kubenswrapper[4944]: I1124 11:20:36.293714 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" path="/var/lib/kubelet/pods/6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5/volumes" Nov 24 11:20:37 crc kubenswrapper[4944]: I1124 11:20:37.993762 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 11:20:37 crc kubenswrapper[4944]: I1124 11:20:37.994125 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 11:20:37 crc kubenswrapper[4944]: I1124 11:20:37.997590 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 11:20:37 crc kubenswrapper[4944]: I1124 11:20:37.997950 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 11:20:38 crc kubenswrapper[4944]: I1124 11:20:38.036897 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 11:20:38 crc kubenswrapper[4944]: I1124 11:20:38.038277 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 11:20:38 crc kubenswrapper[4944]: I1124 11:20:38.038532 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 11:20:38 crc kubenswrapper[4944]: I1124 11:20:38.040297 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 11:20:38 crc kubenswrapper[4944]: I1124 11:20:38.782910 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 11:20:38 crc kubenswrapper[4944]: I1124 11:20:38.785939 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.869764 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp"] Nov 24 11:20:39 crc kubenswrapper[4944]: E1124 11:20:39.870555 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" containerName="extract-content" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.870569 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" containerName="extract-content" Nov 24 11:20:39 crc kubenswrapper[4944]: E1124 11:20:39.870602 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" containerName="registry-server" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.870608 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" containerName="registry-server" Nov 24 11:20:39 crc kubenswrapper[4944]: E1124 11:20:39.870619 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" containerName="extract-utilities" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.870627 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" containerName="extract-utilities" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.870845 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ebf2b3e-cbb0-4b9b-bb3a-6d2d089b38b5" containerName="registry-server" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.871686 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.875631 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.875767 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.876007 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.876024 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.876160 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-crzvv" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.876214 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.878318 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.885376 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp"] Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.977754 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.977822 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.977843 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.978205 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.978538 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2pdr\" (UniqueName: \"kubernetes.io/projected/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-kube-api-access-q2pdr\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.978584 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.978651 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.978671 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.978692 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.978715 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:39 crc kubenswrapper[4944]: I1124 11:20:39.978739 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.079933 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.079991 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.080021 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.080121 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.080169 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2pdr\" (UniqueName: \"kubernetes.io/projected/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-kube-api-access-q2pdr\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.080216 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.080294 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.080317 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.080340 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.080368 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.080403 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.082327 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.083872 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.086741 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.087436 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.087735 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.088644 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.088840 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.088988 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.089021 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.089240 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.106776 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2pdr\" (UniqueName: \"kubernetes.io/projected/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-kube-api-access-q2pdr\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.200608 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.716991 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp"] Nov 24 11:20:40 crc kubenswrapper[4944]: I1124 11:20:40.799657 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" event={"ID":"bf413e07-7450-4eb5-8ff3-3950c9e0d44a","Type":"ContainerStarted","Data":"e5551f144d4e627116108a70a5c971e5757099a0445be61b564467e67a654720"} Nov 24 11:20:42 crc kubenswrapper[4944]: I1124 11:20:42.819311 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" event={"ID":"bf413e07-7450-4eb5-8ff3-3950c9e0d44a","Type":"ContainerStarted","Data":"873118ce9c6156738b8d733b1b6b0ab842e09769a486d3950af46c4b0e4b9fd3"} Nov 24 11:20:42 crc kubenswrapper[4944]: I1124 11:20:42.840836 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" podStartSLOduration=3.461093838 podStartE2EDuration="3.840818589s" podCreationTimestamp="2025-11-24 11:20:39 +0000 UTC" firstStartedPulling="2025-11-24 11:20:40.720018649 +0000 UTC m=+8901.254459111" lastFinishedPulling="2025-11-24 11:20:41.0997434 +0000 UTC m=+8901.634183862" observedRunningTime="2025-11-24 11:20:42.837651579 +0000 UTC m=+8903.372092061" watchObservedRunningTime="2025-11-24 11:20:42.840818589 +0000 UTC m=+8903.375259051" Nov 24 11:20:53 crc kubenswrapper[4944]: I1124 11:20:53.548630 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:20:53 crc kubenswrapper[4944]: I1124 11:20:53.549242 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:21:23 crc kubenswrapper[4944]: I1124 11:21:23.549107 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:21:23 crc kubenswrapper[4944]: I1124 11:21:23.549717 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:21:53 crc kubenswrapper[4944]: I1124 11:21:53.548794 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:21:53 crc kubenswrapper[4944]: I1124 11:21:53.549478 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:21:53 crc kubenswrapper[4944]: I1124 11:21:53.549580 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 11:21:53 crc kubenswrapper[4944]: I1124 11:21:53.550470 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d73d2457794efa51b32f46defdfb5e370f07c447756fe38b7313df14f567eea"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 11:21:53 crc kubenswrapper[4944]: I1124 11:21:53.550542 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://7d73d2457794efa51b32f46defdfb5e370f07c447756fe38b7313df14f567eea" gracePeriod=600 Nov 24 11:21:54 crc kubenswrapper[4944]: I1124 11:21:54.558936 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="7d73d2457794efa51b32f46defdfb5e370f07c447756fe38b7313df14f567eea" exitCode=0 Nov 24 11:21:54 crc kubenswrapper[4944]: I1124 11:21:54.559001 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"7d73d2457794efa51b32f46defdfb5e370f07c447756fe38b7313df14f567eea"} Nov 24 11:21:54 crc kubenswrapper[4944]: I1124 11:21:54.559507 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20"} Nov 24 11:21:54 crc kubenswrapper[4944]: I1124 11:21:54.559545 4944 scope.go:117] "RemoveContainer" containerID="0461e520d39fd81389358bad4dff844aa2749515201e3fea27b73d5a319be95d" Nov 24 11:22:22 crc kubenswrapper[4944]: I1124 11:22:22.780036 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hk9zd"] Nov 24 11:22:22 crc kubenswrapper[4944]: I1124 11:22:22.783909 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:22 crc kubenswrapper[4944]: I1124 11:22:22.796532 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hk9zd"] Nov 24 11:22:22 crc kubenswrapper[4944]: I1124 11:22:22.890673 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k2zb\" (UniqueName: \"kubernetes.io/projected/ed5b24d9-c7e0-426a-bef0-d11e826685d6-kube-api-access-5k2zb\") pod \"certified-operators-hk9zd\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:22 crc kubenswrapper[4944]: I1124 11:22:22.890722 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-catalog-content\") pod \"certified-operators-hk9zd\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:22 crc kubenswrapper[4944]: I1124 11:22:22.890792 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-utilities\") pod \"certified-operators-hk9zd\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:22 crc kubenswrapper[4944]: I1124 11:22:22.993371 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-utilities\") pod \"certified-operators-hk9zd\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:22 crc kubenswrapper[4944]: I1124 11:22:22.993667 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k2zb\" (UniqueName: \"kubernetes.io/projected/ed5b24d9-c7e0-426a-bef0-d11e826685d6-kube-api-access-5k2zb\") pod \"certified-operators-hk9zd\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:22 crc kubenswrapper[4944]: I1124 11:22:22.993712 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-catalog-content\") pod \"certified-operators-hk9zd\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:22 crc kubenswrapper[4944]: I1124 11:22:22.994291 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-catalog-content\") pod \"certified-operators-hk9zd\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:22 crc kubenswrapper[4944]: I1124 11:22:22.994398 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-utilities\") pod \"certified-operators-hk9zd\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:23 crc kubenswrapper[4944]: I1124 11:22:23.028857 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k2zb\" (UniqueName: \"kubernetes.io/projected/ed5b24d9-c7e0-426a-bef0-d11e826685d6-kube-api-access-5k2zb\") pod \"certified-operators-hk9zd\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:23 crc kubenswrapper[4944]: I1124 11:22:23.108227 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:23 crc kubenswrapper[4944]: I1124 11:22:23.631729 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hk9zd"] Nov 24 11:22:23 crc kubenswrapper[4944]: I1124 11:22:23.898851 4944 generic.go:334] "Generic (PLEG): container finished" podID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" containerID="0030465238b9ec194ed957b0d5520674e546047eec08108c2fa1fdb268043462" exitCode=0 Nov 24 11:22:23 crc kubenswrapper[4944]: I1124 11:22:23.898955 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk9zd" event={"ID":"ed5b24d9-c7e0-426a-bef0-d11e826685d6","Type":"ContainerDied","Data":"0030465238b9ec194ed957b0d5520674e546047eec08108c2fa1fdb268043462"} Nov 24 11:22:23 crc kubenswrapper[4944]: I1124 11:22:23.899351 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk9zd" event={"ID":"ed5b24d9-c7e0-426a-bef0-d11e826685d6","Type":"ContainerStarted","Data":"0add53ceea4b8d6bb34167c78b770dfdbc8a11c6782658123c2f7b98be13c98b"} Nov 24 11:22:24 crc kubenswrapper[4944]: I1124 11:22:24.911820 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk9zd" event={"ID":"ed5b24d9-c7e0-426a-bef0-d11e826685d6","Type":"ContainerStarted","Data":"e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668"} Nov 24 11:22:25 crc kubenswrapper[4944]: I1124 11:22:25.923095 4944 generic.go:334] "Generic (PLEG): container finished" podID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" containerID="e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668" exitCode=0 Nov 24 11:22:25 crc kubenswrapper[4944]: I1124 11:22:25.923149 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk9zd" event={"ID":"ed5b24d9-c7e0-426a-bef0-d11e826685d6","Type":"ContainerDied","Data":"e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668"} Nov 24 11:22:26 crc kubenswrapper[4944]: I1124 11:22:26.935712 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk9zd" event={"ID":"ed5b24d9-c7e0-426a-bef0-d11e826685d6","Type":"ContainerStarted","Data":"ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5"} Nov 24 11:22:26 crc kubenswrapper[4944]: I1124 11:22:26.960606 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hk9zd" podStartSLOduration=2.529361636 podStartE2EDuration="4.960584509s" podCreationTimestamp="2025-11-24 11:22:22 +0000 UTC" firstStartedPulling="2025-11-24 11:22:23.901632123 +0000 UTC m=+9004.436072615" lastFinishedPulling="2025-11-24 11:22:26.332855026 +0000 UTC m=+9006.867295488" observedRunningTime="2025-11-24 11:22:26.950347147 +0000 UTC m=+9007.484787609" watchObservedRunningTime="2025-11-24 11:22:26.960584509 +0000 UTC m=+9007.495024981" Nov 24 11:22:33 crc kubenswrapper[4944]: I1124 11:22:33.112409 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:33 crc kubenswrapper[4944]: I1124 11:22:33.112869 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:33 crc kubenswrapper[4944]: I1124 11:22:33.160034 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:34 crc kubenswrapper[4944]: I1124 11:22:34.078442 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:34 crc kubenswrapper[4944]: I1124 11:22:34.129538 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hk9zd"] Nov 24 11:22:36 crc kubenswrapper[4944]: I1124 11:22:36.047913 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hk9zd" podUID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" containerName="registry-server" containerID="cri-o://ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5" gracePeriod=2 Nov 24 11:22:36 crc kubenswrapper[4944]: I1124 11:22:36.551563 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:36 crc kubenswrapper[4944]: I1124 11:22:36.690467 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k2zb\" (UniqueName: \"kubernetes.io/projected/ed5b24d9-c7e0-426a-bef0-d11e826685d6-kube-api-access-5k2zb\") pod \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " Nov 24 11:22:36 crc kubenswrapper[4944]: I1124 11:22:36.690531 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-catalog-content\") pod \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " Nov 24 11:22:36 crc kubenswrapper[4944]: I1124 11:22:36.690682 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-utilities\") pod \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\" (UID: \"ed5b24d9-c7e0-426a-bef0-d11e826685d6\") " Nov 24 11:22:36 crc kubenswrapper[4944]: I1124 11:22:36.692156 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-utilities" (OuterVolumeSpecName: "utilities") pod "ed5b24d9-c7e0-426a-bef0-d11e826685d6" (UID: "ed5b24d9-c7e0-426a-bef0-d11e826685d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:22:36 crc kubenswrapper[4944]: I1124 11:22:36.697016 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed5b24d9-c7e0-426a-bef0-d11e826685d6-kube-api-access-5k2zb" (OuterVolumeSpecName: "kube-api-access-5k2zb") pod "ed5b24d9-c7e0-426a-bef0-d11e826685d6" (UID: "ed5b24d9-c7e0-426a-bef0-d11e826685d6"). InnerVolumeSpecName "kube-api-access-5k2zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:22:36 crc kubenswrapper[4944]: I1124 11:22:36.745121 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed5b24d9-c7e0-426a-bef0-d11e826685d6" (UID: "ed5b24d9-c7e0-426a-bef0-d11e826685d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:22:36 crc kubenswrapper[4944]: I1124 11:22:36.793659 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k2zb\" (UniqueName: \"kubernetes.io/projected/ed5b24d9-c7e0-426a-bef0-d11e826685d6-kube-api-access-5k2zb\") on node \"crc\" DevicePath \"\"" Nov 24 11:22:36 crc kubenswrapper[4944]: I1124 11:22:36.793903 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:22:36 crc kubenswrapper[4944]: I1124 11:22:36.793971 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed5b24d9-c7e0-426a-bef0-d11e826685d6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.058098 4944 generic.go:334] "Generic (PLEG): container finished" podID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" containerID="ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5" exitCode=0 Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.058152 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk9zd" event={"ID":"ed5b24d9-c7e0-426a-bef0-d11e826685d6","Type":"ContainerDied","Data":"ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5"} Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.058211 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk9zd" event={"ID":"ed5b24d9-c7e0-426a-bef0-d11e826685d6","Type":"ContainerDied","Data":"0add53ceea4b8d6bb34167c78b770dfdbc8a11c6782658123c2f7b98be13c98b"} Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.058230 4944 scope.go:117] "RemoveContainer" containerID="ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5" Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.059130 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hk9zd" Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.077732 4944 scope.go:117] "RemoveContainer" containerID="e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668" Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.100229 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hk9zd"] Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.112010 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hk9zd"] Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.120262 4944 scope.go:117] "RemoveContainer" containerID="0030465238b9ec194ed957b0d5520674e546047eec08108c2fa1fdb268043462" Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.148155 4944 scope.go:117] "RemoveContainer" containerID="ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5" Nov 24 11:22:37 crc kubenswrapper[4944]: E1124 11:22:37.148561 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5\": container with ID starting with ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5 not found: ID does not exist" containerID="ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5" Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.148594 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5"} err="failed to get container status \"ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5\": rpc error: code = NotFound desc = could not find container \"ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5\": container with ID starting with ebc702be8e7770f891078bbe3cbb0e751c4e4b3d8cf9dda93d8bb935903e97b5 not found: ID does not exist" Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.148613 4944 scope.go:117] "RemoveContainer" containerID="e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668" Nov 24 11:22:37 crc kubenswrapper[4944]: E1124 11:22:37.148884 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668\": container with ID starting with e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668 not found: ID does not exist" containerID="e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668" Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.148913 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668"} err="failed to get container status \"e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668\": rpc error: code = NotFound desc = could not find container \"e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668\": container with ID starting with e9bdd1b3e6ec8d0035ebb2615746d4bee2b4009b203da2f4a7d1f5ad35791668 not found: ID does not exist" Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.148927 4944 scope.go:117] "RemoveContainer" containerID="0030465238b9ec194ed957b0d5520674e546047eec08108c2fa1fdb268043462" Nov 24 11:22:37 crc kubenswrapper[4944]: E1124 11:22:37.149186 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0030465238b9ec194ed957b0d5520674e546047eec08108c2fa1fdb268043462\": container with ID starting with 0030465238b9ec194ed957b0d5520674e546047eec08108c2fa1fdb268043462 not found: ID does not exist" containerID="0030465238b9ec194ed957b0d5520674e546047eec08108c2fa1fdb268043462" Nov 24 11:22:37 crc kubenswrapper[4944]: I1124 11:22:37.149209 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0030465238b9ec194ed957b0d5520674e546047eec08108c2fa1fdb268043462"} err="failed to get container status \"0030465238b9ec194ed957b0d5520674e546047eec08108c2fa1fdb268043462\": rpc error: code = NotFound desc = could not find container \"0030465238b9ec194ed957b0d5520674e546047eec08108c2fa1fdb268043462\": container with ID starting with 0030465238b9ec194ed957b0d5520674e546047eec08108c2fa1fdb268043462 not found: ID does not exist" Nov 24 11:22:38 crc kubenswrapper[4944]: I1124 11:22:38.287879 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" path="/var/lib/kubelet/pods/ed5b24d9-c7e0-426a-bef0-d11e826685d6/volumes" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.092897 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f22jr"] Nov 24 11:23:50 crc kubenswrapper[4944]: E1124 11:23:50.093857 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" containerName="registry-server" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.093875 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" containerName="registry-server" Nov 24 11:23:50 crc kubenswrapper[4944]: E1124 11:23:50.093886 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" containerName="extract-content" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.093892 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" containerName="extract-content" Nov 24 11:23:50 crc kubenswrapper[4944]: E1124 11:23:50.093935 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" containerName="extract-utilities" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.093941 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" containerName="extract-utilities" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.094158 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed5b24d9-c7e0-426a-bef0-d11e826685d6" containerName="registry-server" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.095789 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.102924 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f22jr"] Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.249894 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-utilities\") pod \"community-operators-f22jr\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.249988 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvk6s\" (UniqueName: \"kubernetes.io/projected/2425953b-57c0-4d80-b41c-c6af84bfa1bf-kube-api-access-kvk6s\") pod \"community-operators-f22jr\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.250041 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-catalog-content\") pod \"community-operators-f22jr\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.351987 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-utilities\") pod \"community-operators-f22jr\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.352135 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvk6s\" (UniqueName: \"kubernetes.io/projected/2425953b-57c0-4d80-b41c-c6af84bfa1bf-kube-api-access-kvk6s\") pod \"community-operators-f22jr\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.352208 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-catalog-content\") pod \"community-operators-f22jr\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.352994 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-utilities\") pod \"community-operators-f22jr\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.353129 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-catalog-content\") pod \"community-operators-f22jr\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.697284 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvk6s\" (UniqueName: \"kubernetes.io/projected/2425953b-57c0-4d80-b41c-c6af84bfa1bf-kube-api-access-kvk6s\") pod \"community-operators-f22jr\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:23:50 crc kubenswrapper[4944]: I1124 11:23:50.718177 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:23:51 crc kubenswrapper[4944]: I1124 11:23:51.161664 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f22jr"] Nov 24 11:23:51 crc kubenswrapper[4944]: I1124 11:23:51.845578 4944 generic.go:334] "Generic (PLEG): container finished" podID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" containerID="f5391ad5cc803bef854fe6383419d0b55fd8f949b609680b38ced773b32b46f9" exitCode=0 Nov 24 11:23:51 crc kubenswrapper[4944]: I1124 11:23:51.845634 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f22jr" event={"ID":"2425953b-57c0-4d80-b41c-c6af84bfa1bf","Type":"ContainerDied","Data":"f5391ad5cc803bef854fe6383419d0b55fd8f949b609680b38ced773b32b46f9"} Nov 24 11:23:51 crc kubenswrapper[4944]: I1124 11:23:51.845929 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f22jr" event={"ID":"2425953b-57c0-4d80-b41c-c6af84bfa1bf","Type":"ContainerStarted","Data":"e00b0fef9651e788698464d8d6b936d760757dec278c8a806cea69fbb2ed5ca6"} Nov 24 11:23:52 crc kubenswrapper[4944]: I1124 11:23:52.858539 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f22jr" event={"ID":"2425953b-57c0-4d80-b41c-c6af84bfa1bf","Type":"ContainerStarted","Data":"68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b"} Nov 24 11:23:53 crc kubenswrapper[4944]: I1124 11:23:53.548178 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:23:53 crc kubenswrapper[4944]: I1124 11:23:53.548534 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:23:53 crc kubenswrapper[4944]: I1124 11:23:53.870491 4944 generic.go:334] "Generic (PLEG): container finished" podID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" containerID="68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b" exitCode=0 Nov 24 11:23:53 crc kubenswrapper[4944]: I1124 11:23:53.870558 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f22jr" event={"ID":"2425953b-57c0-4d80-b41c-c6af84bfa1bf","Type":"ContainerDied","Data":"68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b"} Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.329681 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mpflr"] Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.332236 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mpflr"] Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.332333 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.452311 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-catalog-content\") pod \"redhat-operators-mpflr\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.452377 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-utilities\") pod \"redhat-operators-mpflr\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.452452 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr6rk\" (UniqueName: \"kubernetes.io/projected/66850c80-f50b-413d-aec6-2dfcc9d3a633-kube-api-access-lr6rk\") pod \"redhat-operators-mpflr\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.554393 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr6rk\" (UniqueName: \"kubernetes.io/projected/66850c80-f50b-413d-aec6-2dfcc9d3a633-kube-api-access-lr6rk\") pod \"redhat-operators-mpflr\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.554702 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-catalog-content\") pod \"redhat-operators-mpflr\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.554742 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-utilities\") pod \"redhat-operators-mpflr\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.555887 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-utilities\") pod \"redhat-operators-mpflr\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.556003 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-catalog-content\") pod \"redhat-operators-mpflr\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.579786 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr6rk\" (UniqueName: \"kubernetes.io/projected/66850c80-f50b-413d-aec6-2dfcc9d3a633-kube-api-access-lr6rk\") pod \"redhat-operators-mpflr\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.680676 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:23:54 crc kubenswrapper[4944]: I1124 11:23:54.950380 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f22jr" event={"ID":"2425953b-57c0-4d80-b41c-c6af84bfa1bf","Type":"ContainerStarted","Data":"07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7"} Nov 24 11:23:55 crc kubenswrapper[4944]: I1124 11:23:55.000634 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f22jr" podStartSLOduration=2.527420486 podStartE2EDuration="5.000615033s" podCreationTimestamp="2025-11-24 11:23:50 +0000 UTC" firstStartedPulling="2025-11-24 11:23:51.847243737 +0000 UTC m=+9092.381684199" lastFinishedPulling="2025-11-24 11:23:54.320438284 +0000 UTC m=+9094.854878746" observedRunningTime="2025-11-24 11:23:54.992662152 +0000 UTC m=+9095.527102614" watchObservedRunningTime="2025-11-24 11:23:55.000615033 +0000 UTC m=+9095.535055495" Nov 24 11:23:55 crc kubenswrapper[4944]: I1124 11:23:55.229440 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mpflr"] Nov 24 11:23:55 crc kubenswrapper[4944]: W1124 11:23:55.232094 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66850c80_f50b_413d_aec6_2dfcc9d3a633.slice/crio-fccfc636baaf9fd48ad3a35efacf9ea672927db337d5a85457a1e5c4d81e67a3 WatchSource:0}: Error finding container fccfc636baaf9fd48ad3a35efacf9ea672927db337d5a85457a1e5c4d81e67a3: Status 404 returned error can't find the container with id fccfc636baaf9fd48ad3a35efacf9ea672927db337d5a85457a1e5c4d81e67a3 Nov 24 11:23:55 crc kubenswrapper[4944]: I1124 11:23:55.961828 4944 generic.go:334] "Generic (PLEG): container finished" podID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerID="9dc71bf2dc3ab9527fc8e2f40bc01ca78025552ddf8be33d6d1549f4fdc825d6" exitCode=0 Nov 24 11:23:55 crc kubenswrapper[4944]: I1124 11:23:55.961883 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpflr" event={"ID":"66850c80-f50b-413d-aec6-2dfcc9d3a633","Type":"ContainerDied","Data":"9dc71bf2dc3ab9527fc8e2f40bc01ca78025552ddf8be33d6d1549f4fdc825d6"} Nov 24 11:23:55 crc kubenswrapper[4944]: I1124 11:23:55.962331 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpflr" event={"ID":"66850c80-f50b-413d-aec6-2dfcc9d3a633","Type":"ContainerStarted","Data":"fccfc636baaf9fd48ad3a35efacf9ea672927db337d5a85457a1e5c4d81e67a3"} Nov 24 11:23:56 crc kubenswrapper[4944]: I1124 11:23:56.974378 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpflr" event={"ID":"66850c80-f50b-413d-aec6-2dfcc9d3a633","Type":"ContainerStarted","Data":"a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e"} Nov 24 11:24:00 crc kubenswrapper[4944]: I1124 11:24:00.009550 4944 generic.go:334] "Generic (PLEG): container finished" podID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerID="a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e" exitCode=0 Nov 24 11:24:00 crc kubenswrapper[4944]: I1124 11:24:00.009632 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpflr" event={"ID":"66850c80-f50b-413d-aec6-2dfcc9d3a633","Type":"ContainerDied","Data":"a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e"} Nov 24 11:24:00 crc kubenswrapper[4944]: I1124 11:24:00.718288 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:24:00 crc kubenswrapper[4944]: I1124 11:24:00.718596 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:24:00 crc kubenswrapper[4944]: I1124 11:24:00.769868 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:24:01 crc kubenswrapper[4944]: I1124 11:24:01.021032 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpflr" event={"ID":"66850c80-f50b-413d-aec6-2dfcc9d3a633","Type":"ContainerStarted","Data":"5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1"} Nov 24 11:24:01 crc kubenswrapper[4944]: I1124 11:24:01.038864 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mpflr" podStartSLOduration=2.502468554 podStartE2EDuration="7.038850773s" podCreationTimestamp="2025-11-24 11:23:54 +0000 UTC" firstStartedPulling="2025-11-24 11:23:55.965903436 +0000 UTC m=+9096.500343898" lastFinishedPulling="2025-11-24 11:24:00.502285655 +0000 UTC m=+9101.036726117" observedRunningTime="2025-11-24 11:24:01.037940474 +0000 UTC m=+9101.572380936" watchObservedRunningTime="2025-11-24 11:24:01.038850773 +0000 UTC m=+9101.573291235" Nov 24 11:24:01 crc kubenswrapper[4944]: I1124 11:24:01.085709 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:24:02 crc kubenswrapper[4944]: I1124 11:24:02.267265 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f22jr"] Nov 24 11:24:03 crc kubenswrapper[4944]: I1124 11:24:03.036038 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f22jr" podUID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" containerName="registry-server" containerID="cri-o://07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7" gracePeriod=2 Nov 24 11:24:03 crc kubenswrapper[4944]: I1124 11:24:03.567527 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:24:03 crc kubenswrapper[4944]: I1124 11:24:03.677781 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-utilities\") pod \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " Nov 24 11:24:03 crc kubenswrapper[4944]: I1124 11:24:03.677951 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-catalog-content\") pod \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " Nov 24 11:24:03 crc kubenswrapper[4944]: I1124 11:24:03.678009 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvk6s\" (UniqueName: \"kubernetes.io/projected/2425953b-57c0-4d80-b41c-c6af84bfa1bf-kube-api-access-kvk6s\") pod \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\" (UID: \"2425953b-57c0-4d80-b41c-c6af84bfa1bf\") " Nov 24 11:24:03 crc kubenswrapper[4944]: I1124 11:24:03.679870 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-utilities" (OuterVolumeSpecName: "utilities") pod "2425953b-57c0-4d80-b41c-c6af84bfa1bf" (UID: "2425953b-57c0-4d80-b41c-c6af84bfa1bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:24:03 crc kubenswrapper[4944]: I1124 11:24:03.684146 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2425953b-57c0-4d80-b41c-c6af84bfa1bf-kube-api-access-kvk6s" (OuterVolumeSpecName: "kube-api-access-kvk6s") pod "2425953b-57c0-4d80-b41c-c6af84bfa1bf" (UID: "2425953b-57c0-4d80-b41c-c6af84bfa1bf"). InnerVolumeSpecName "kube-api-access-kvk6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:24:03 crc kubenswrapper[4944]: I1124 11:24:03.745556 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2425953b-57c0-4d80-b41c-c6af84bfa1bf" (UID: "2425953b-57c0-4d80-b41c-c6af84bfa1bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:24:03 crc kubenswrapper[4944]: I1124 11:24:03.780382 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:24:03 crc kubenswrapper[4944]: I1124 11:24:03.780426 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvk6s\" (UniqueName: \"kubernetes.io/projected/2425953b-57c0-4d80-b41c-c6af84bfa1bf-kube-api-access-kvk6s\") on node \"crc\" DevicePath \"\"" Nov 24 11:24:03 crc kubenswrapper[4944]: I1124 11:24:03.780442 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2425953b-57c0-4d80-b41c-c6af84bfa1bf-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.046935 4944 generic.go:334] "Generic (PLEG): container finished" podID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" containerID="07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7" exitCode=0 Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.046981 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f22jr" event={"ID":"2425953b-57c0-4d80-b41c-c6af84bfa1bf","Type":"ContainerDied","Data":"07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7"} Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.047007 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f22jr" event={"ID":"2425953b-57c0-4d80-b41c-c6af84bfa1bf","Type":"ContainerDied","Data":"e00b0fef9651e788698464d8d6b936d760757dec278c8a806cea69fbb2ed5ca6"} Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.047023 4944 scope.go:117] "RemoveContainer" containerID="07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.047188 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f22jr" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.076693 4944 scope.go:117] "RemoveContainer" containerID="68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.080512 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f22jr"] Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.094042 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f22jr"] Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.121527 4944 scope.go:117] "RemoveContainer" containerID="f5391ad5cc803bef854fe6383419d0b55fd8f949b609680b38ced773b32b46f9" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.156422 4944 scope.go:117] "RemoveContainer" containerID="07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7" Nov 24 11:24:04 crc kubenswrapper[4944]: E1124 11:24:04.156887 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7\": container with ID starting with 07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7 not found: ID does not exist" containerID="07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.156920 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7"} err="failed to get container status \"07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7\": rpc error: code = NotFound desc = could not find container \"07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7\": container with ID starting with 07861d128c62a082868f147c8c13d860c22d217ffaab8dac30d3b0696b055bb7 not found: ID does not exist" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.156947 4944 scope.go:117] "RemoveContainer" containerID="68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b" Nov 24 11:24:04 crc kubenswrapper[4944]: E1124 11:24:04.157197 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b\": container with ID starting with 68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b not found: ID does not exist" containerID="68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.157221 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b"} err="failed to get container status \"68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b\": rpc error: code = NotFound desc = could not find container \"68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b\": container with ID starting with 68989217137cb2537b032f3554b20a1a35240bcd92d4754933f0d80a29ca260b not found: ID does not exist" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.157236 4944 scope.go:117] "RemoveContainer" containerID="f5391ad5cc803bef854fe6383419d0b55fd8f949b609680b38ced773b32b46f9" Nov 24 11:24:04 crc kubenswrapper[4944]: E1124 11:24:04.157461 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5391ad5cc803bef854fe6383419d0b55fd8f949b609680b38ced773b32b46f9\": container with ID starting with f5391ad5cc803bef854fe6383419d0b55fd8f949b609680b38ced773b32b46f9 not found: ID does not exist" containerID="f5391ad5cc803bef854fe6383419d0b55fd8f949b609680b38ced773b32b46f9" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.157484 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5391ad5cc803bef854fe6383419d0b55fd8f949b609680b38ced773b32b46f9"} err="failed to get container status \"f5391ad5cc803bef854fe6383419d0b55fd8f949b609680b38ced773b32b46f9\": rpc error: code = NotFound desc = could not find container \"f5391ad5cc803bef854fe6383419d0b55fd8f949b609680b38ced773b32b46f9\": container with ID starting with f5391ad5cc803bef854fe6383419d0b55fd8f949b609680b38ced773b32b46f9 not found: ID does not exist" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.291253 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" path="/var/lib/kubelet/pods/2425953b-57c0-4d80-b41c-c6af84bfa1bf/volumes" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.681666 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:24:04 crc kubenswrapper[4944]: I1124 11:24:04.681713 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:24:05 crc kubenswrapper[4944]: I1124 11:24:05.728225 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mpflr" podUID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerName="registry-server" probeResult="failure" output=< Nov 24 11:24:05 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 11:24:05 crc kubenswrapper[4944]: > Nov 24 11:24:14 crc kubenswrapper[4944]: I1124 11:24:14.753959 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:24:14 crc kubenswrapper[4944]: I1124 11:24:14.822999 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:24:14 crc kubenswrapper[4944]: I1124 11:24:14.998670 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mpflr"] Nov 24 11:24:16 crc kubenswrapper[4944]: I1124 11:24:16.158562 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mpflr" podUID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerName="registry-server" containerID="cri-o://5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1" gracePeriod=2 Nov 24 11:24:16 crc kubenswrapper[4944]: I1124 11:24:16.670880 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:24:16 crc kubenswrapper[4944]: I1124 11:24:16.802603 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-utilities\") pod \"66850c80-f50b-413d-aec6-2dfcc9d3a633\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " Nov 24 11:24:16 crc kubenswrapper[4944]: I1124 11:24:16.802729 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr6rk\" (UniqueName: \"kubernetes.io/projected/66850c80-f50b-413d-aec6-2dfcc9d3a633-kube-api-access-lr6rk\") pod \"66850c80-f50b-413d-aec6-2dfcc9d3a633\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " Nov 24 11:24:16 crc kubenswrapper[4944]: I1124 11:24:16.802817 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-catalog-content\") pod \"66850c80-f50b-413d-aec6-2dfcc9d3a633\" (UID: \"66850c80-f50b-413d-aec6-2dfcc9d3a633\") " Nov 24 11:24:16 crc kubenswrapper[4944]: I1124 11:24:16.803617 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-utilities" (OuterVolumeSpecName: "utilities") pod "66850c80-f50b-413d-aec6-2dfcc9d3a633" (UID: "66850c80-f50b-413d-aec6-2dfcc9d3a633"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:24:16 crc kubenswrapper[4944]: I1124 11:24:16.808434 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66850c80-f50b-413d-aec6-2dfcc9d3a633-kube-api-access-lr6rk" (OuterVolumeSpecName: "kube-api-access-lr6rk") pod "66850c80-f50b-413d-aec6-2dfcc9d3a633" (UID: "66850c80-f50b-413d-aec6-2dfcc9d3a633"). InnerVolumeSpecName "kube-api-access-lr6rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:24:16 crc kubenswrapper[4944]: I1124 11:24:16.896676 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66850c80-f50b-413d-aec6-2dfcc9d3a633" (UID: "66850c80-f50b-413d-aec6-2dfcc9d3a633"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:24:16 crc kubenswrapper[4944]: I1124 11:24:16.905065 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:24:16 crc kubenswrapper[4944]: I1124 11:24:16.905124 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr6rk\" (UniqueName: \"kubernetes.io/projected/66850c80-f50b-413d-aec6-2dfcc9d3a633-kube-api-access-lr6rk\") on node \"crc\" DevicePath \"\"" Nov 24 11:24:16 crc kubenswrapper[4944]: I1124 11:24:16.905140 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66850c80-f50b-413d-aec6-2dfcc9d3a633-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.169822 4944 generic.go:334] "Generic (PLEG): container finished" podID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerID="5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1" exitCode=0 Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.169872 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpflr" Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.169877 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpflr" event={"ID":"66850c80-f50b-413d-aec6-2dfcc9d3a633","Type":"ContainerDied","Data":"5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1"} Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.170016 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpflr" event={"ID":"66850c80-f50b-413d-aec6-2dfcc9d3a633","Type":"ContainerDied","Data":"fccfc636baaf9fd48ad3a35efacf9ea672927db337d5a85457a1e5c4d81e67a3"} Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.170067 4944 scope.go:117] "RemoveContainer" containerID="5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1" Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.189916 4944 scope.go:117] "RemoveContainer" containerID="a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e" Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.205802 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mpflr"] Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.215585 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mpflr"] Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.227734 4944 scope.go:117] "RemoveContainer" containerID="9dc71bf2dc3ab9527fc8e2f40bc01ca78025552ddf8be33d6d1549f4fdc825d6" Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.274209 4944 scope.go:117] "RemoveContainer" containerID="5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1" Nov 24 11:24:17 crc kubenswrapper[4944]: E1124 11:24:17.275233 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1\": container with ID starting with 5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1 not found: ID does not exist" containerID="5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1" Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.275269 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1"} err="failed to get container status \"5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1\": rpc error: code = NotFound desc = could not find container \"5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1\": container with ID starting with 5a80915c302513e8910095d21b3dc962e051b666d1f749189937609d66e2dcb1 not found: ID does not exist" Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.275296 4944 scope.go:117] "RemoveContainer" containerID="a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e" Nov 24 11:24:17 crc kubenswrapper[4944]: E1124 11:24:17.275713 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e\": container with ID starting with a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e not found: ID does not exist" containerID="a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e" Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.276012 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e"} err="failed to get container status \"a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e\": rpc error: code = NotFound desc = could not find container \"a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e\": container with ID starting with a0bd1397c5700515bdddb3f6a1b8b438fb78e7815e2aa609a487cf749804c98e not found: ID does not exist" Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.276215 4944 scope.go:117] "RemoveContainer" containerID="9dc71bf2dc3ab9527fc8e2f40bc01ca78025552ddf8be33d6d1549f4fdc825d6" Nov 24 11:24:17 crc kubenswrapper[4944]: E1124 11:24:17.276699 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dc71bf2dc3ab9527fc8e2f40bc01ca78025552ddf8be33d6d1549f4fdc825d6\": container with ID starting with 9dc71bf2dc3ab9527fc8e2f40bc01ca78025552ddf8be33d6d1549f4fdc825d6 not found: ID does not exist" containerID="9dc71bf2dc3ab9527fc8e2f40bc01ca78025552ddf8be33d6d1549f4fdc825d6" Nov 24 11:24:17 crc kubenswrapper[4944]: I1124 11:24:17.276744 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dc71bf2dc3ab9527fc8e2f40bc01ca78025552ddf8be33d6d1549f4fdc825d6"} err="failed to get container status \"9dc71bf2dc3ab9527fc8e2f40bc01ca78025552ddf8be33d6d1549f4fdc825d6\": rpc error: code = NotFound desc = could not find container \"9dc71bf2dc3ab9527fc8e2f40bc01ca78025552ddf8be33d6d1549f4fdc825d6\": container with ID starting with 9dc71bf2dc3ab9527fc8e2f40bc01ca78025552ddf8be33d6d1549f4fdc825d6 not found: ID does not exist" Nov 24 11:24:18 crc kubenswrapper[4944]: I1124 11:24:18.288642 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66850c80-f50b-413d-aec6-2dfcc9d3a633" path="/var/lib/kubelet/pods/66850c80-f50b-413d-aec6-2dfcc9d3a633/volumes" Nov 24 11:24:23 crc kubenswrapper[4944]: I1124 11:24:23.548595 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:24:23 crc kubenswrapper[4944]: I1124 11:24:23.549090 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:24:53 crc kubenswrapper[4944]: I1124 11:24:53.548960 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:24:53 crc kubenswrapper[4944]: I1124 11:24:53.550164 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:24:53 crc kubenswrapper[4944]: I1124 11:24:53.550251 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 11:24:53 crc kubenswrapper[4944]: I1124 11:24:53.551727 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 11:24:53 crc kubenswrapper[4944]: I1124 11:24:53.551850 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" gracePeriod=600 Nov 24 11:24:54 crc kubenswrapper[4944]: E1124 11:24:54.181216 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:24:54 crc kubenswrapper[4944]: I1124 11:24:54.556258 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" exitCode=0 Nov 24 11:24:54 crc kubenswrapper[4944]: I1124 11:24:54.556303 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20"} Nov 24 11:24:54 crc kubenswrapper[4944]: I1124 11:24:54.556336 4944 scope.go:117] "RemoveContainer" containerID="7d73d2457794efa51b32f46defdfb5e370f07c447756fe38b7313df14f567eea" Nov 24 11:24:54 crc kubenswrapper[4944]: I1124 11:24:54.557340 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:24:54 crc kubenswrapper[4944]: E1124 11:24:54.557916 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:25:06 crc kubenswrapper[4944]: I1124 11:25:06.277850 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:25:06 crc kubenswrapper[4944]: E1124 11:25:06.278834 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:25:19 crc kubenswrapper[4944]: I1124 11:25:19.277541 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:25:19 crc kubenswrapper[4944]: E1124 11:25:19.278540 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:25:30 crc kubenswrapper[4944]: I1124 11:25:30.287086 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:25:30 crc kubenswrapper[4944]: E1124 11:25:30.288177 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:25:45 crc kubenswrapper[4944]: I1124 11:25:45.277424 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:25:45 crc kubenswrapper[4944]: E1124 11:25:45.278227 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:25:58 crc kubenswrapper[4944]: I1124 11:25:58.277250 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:25:58 crc kubenswrapper[4944]: E1124 11:25:58.278379 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:26:12 crc kubenswrapper[4944]: I1124 11:26:12.276506 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:26:12 crc kubenswrapper[4944]: E1124 11:26:12.277477 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:26:12 crc kubenswrapper[4944]: I1124 11:26:12.366933 4944 generic.go:334] "Generic (PLEG): container finished" podID="bf413e07-7450-4eb5-8ff3-3950c9e0d44a" containerID="873118ce9c6156738b8d733b1b6b0ab842e09769a486d3950af46c4b0e4b9fd3" exitCode=0 Nov 24 11:26:12 crc kubenswrapper[4944]: I1124 11:26:12.366977 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" event={"ID":"bf413e07-7450-4eb5-8ff3-3950c9e0d44a","Type":"ContainerDied","Data":"873118ce9c6156738b8d733b1b6b0ab842e09769a486d3950af46c4b0e4b9fd3"} Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.838287 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.933859 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-inventory\") pod \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.933977 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ssh-key\") pod \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.934038 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-0\") pod \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.934105 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-0\") pod \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.934143 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-1\") pod \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.934178 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-1\") pod \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.934228 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-combined-ca-bundle\") pod \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.934268 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ceph\") pod \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.934318 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2pdr\" (UniqueName: \"kubernetes.io/projected/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-kube-api-access-q2pdr\") pod \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.934356 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-1\") pod \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.934394 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-0\") pod \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\" (UID: \"bf413e07-7450-4eb5-8ff3-3950c9e0d44a\") " Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.940147 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-kube-api-access-q2pdr" (OuterVolumeSpecName: "kube-api-access-q2pdr") pod "bf413e07-7450-4eb5-8ff3-3950c9e0d44a" (UID: "bf413e07-7450-4eb5-8ff3-3950c9e0d44a"). InnerVolumeSpecName "kube-api-access-q2pdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.960551 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ceph" (OuterVolumeSpecName: "ceph") pod "bf413e07-7450-4eb5-8ff3-3950c9e0d44a" (UID: "bf413e07-7450-4eb5-8ff3-3950c9e0d44a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.961276 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "bf413e07-7450-4eb5-8ff3-3950c9e0d44a" (UID: "bf413e07-7450-4eb5-8ff3-3950c9e0d44a"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.964595 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "bf413e07-7450-4eb5-8ff3-3950c9e0d44a" (UID: "bf413e07-7450-4eb5-8ff3-3950c9e0d44a"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.965174 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "bf413e07-7450-4eb5-8ff3-3950c9e0d44a" (UID: "bf413e07-7450-4eb5-8ff3-3950c9e0d44a"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.965538 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "bf413e07-7450-4eb5-8ff3-3950c9e0d44a" (UID: "bf413e07-7450-4eb5-8ff3-3950c9e0d44a"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.967352 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "bf413e07-7450-4eb5-8ff3-3950c9e0d44a" (UID: "bf413e07-7450-4eb5-8ff3-3950c9e0d44a"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.983936 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-inventory" (OuterVolumeSpecName: "inventory") pod "bf413e07-7450-4eb5-8ff3-3950c9e0d44a" (UID: "bf413e07-7450-4eb5-8ff3-3950c9e0d44a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.984183 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf413e07-7450-4eb5-8ff3-3950c9e0d44a" (UID: "bf413e07-7450-4eb5-8ff3-3950c9e0d44a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.993388 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "bf413e07-7450-4eb5-8ff3-3950c9e0d44a" (UID: "bf413e07-7450-4eb5-8ff3-3950c9e0d44a"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 11:26:13 crc kubenswrapper[4944]: I1124 11:26:13.994932 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "bf413e07-7450-4eb5-8ff3-3950c9e0d44a" (UID: "bf413e07-7450-4eb5-8ff3-3950c9e0d44a"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.036336 4944 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.036376 4944 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.036391 4944 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.036404 4944 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.036416 4944 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.036428 4944 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.036438 4944 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.036449 4944 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-ceph\") on node \"crc\" DevicePath \"\"" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.036460 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2pdr\" (UniqueName: \"kubernetes.io/projected/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-kube-api-access-q2pdr\") on node \"crc\" DevicePath \"\"" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.036472 4944 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.036482 4944 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bf413e07-7450-4eb5-8ff3-3950c9e0d44a-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.392628 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" event={"ID":"bf413e07-7450-4eb5-8ff3-3950c9e0d44a","Type":"ContainerDied","Data":"e5551f144d4e627116108a70a5c971e5757099a0445be61b564467e67a654720"} Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.392676 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5551f144d4e627116108a70a5c971e5757099a0445be61b564467e67a654720" Nov 24 11:26:14 crc kubenswrapper[4944]: I1124 11:26:14.392798 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp" Nov 24 11:26:24 crc kubenswrapper[4944]: I1124 11:26:24.279226 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:26:24 crc kubenswrapper[4944]: E1124 11:26:24.280262 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:26:36 crc kubenswrapper[4944]: I1124 11:26:36.531460 4944 trace.go:236] Trace[1285954834]: "Calculate volume metrics of prometheus-metric-storage-db for pod openstack/prometheus-metric-storage-0" (24-Nov-2025 11:26:35.312) (total time: 1218ms): Nov 24 11:26:36 crc kubenswrapper[4944]: Trace[1285954834]: [1.218550104s] [1.218550104s] END Nov 24 11:26:39 crc kubenswrapper[4944]: I1124 11:26:39.277148 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:26:39 crc kubenswrapper[4944]: E1124 11:26:39.277650 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:26:54 crc kubenswrapper[4944]: I1124 11:26:54.278905 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:26:54 crc kubenswrapper[4944]: E1124 11:26:54.280388 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:27:06 crc kubenswrapper[4944]: I1124 11:27:06.277329 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:27:06 crc kubenswrapper[4944]: E1124 11:27:06.278101 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:27:20 crc kubenswrapper[4944]: I1124 11:27:20.289357 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:27:20 crc kubenswrapper[4944]: E1124 11:27:20.290789 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:27:31 crc kubenswrapper[4944]: I1124 11:27:31.796140 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="56ccc568-803f-4862-b9d6-f5e29d45f5b0" containerName="galera" probeResult="failure" output="command timed out" Nov 24 11:27:31 crc kubenswrapper[4944]: I1124 11:27:31.797099 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="56ccc568-803f-4862-b9d6-f5e29d45f5b0" containerName="galera" probeResult="failure" output="command timed out" Nov 24 11:27:34 crc kubenswrapper[4944]: I1124 11:27:34.282064 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:27:34 crc kubenswrapper[4944]: E1124 11:27:34.282824 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:27:49 crc kubenswrapper[4944]: I1124 11:27:49.276884 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:27:49 crc kubenswrapper[4944]: E1124 11:27:49.277582 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:28:02 crc kubenswrapper[4944]: I1124 11:28:02.278124 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:28:02 crc kubenswrapper[4944]: E1124 11:28:02.278974 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:28:07 crc kubenswrapper[4944]: E1124 11:28:07.987327 4944 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.15:36666->38.129.56.15:34895: write tcp 38.129.56.15:36666->38.129.56.15:34895: write: broken pipe Nov 24 11:28:14 crc kubenswrapper[4944]: I1124 11:28:14.280788 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:28:14 crc kubenswrapper[4944]: E1124 11:28:14.281627 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:28:23 crc kubenswrapper[4944]: I1124 11:28:23.944962 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 11:28:23 crc kubenswrapper[4944]: I1124 11:28:23.945765 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="9ff0f1b1-a485-45a1-8a01-e59eb360445d" containerName="adoption" containerID="cri-o://4c97f4fa04a0fb4a629e12fd1304cd09082f79eaebd148be0cdabcea44912e02" gracePeriod=30 Nov 24 11:28:28 crc kubenswrapper[4944]: I1124 11:28:28.277401 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:28:28 crc kubenswrapper[4944]: E1124 11:28:28.278393 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:28:41 crc kubenswrapper[4944]: I1124 11:28:41.277952 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:28:41 crc kubenswrapper[4944]: E1124 11:28:41.279378 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.119492 4944 generic.go:334] "Generic (PLEG): container finished" podID="9ff0f1b1-a485-45a1-8a01-e59eb360445d" containerID="4c97f4fa04a0fb4a629e12fd1304cd09082f79eaebd148be0cdabcea44912e02" exitCode=137 Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.120193 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"9ff0f1b1-a485-45a1-8a01-e59eb360445d","Type":"ContainerDied","Data":"4c97f4fa04a0fb4a629e12fd1304cd09082f79eaebd148be0cdabcea44912e02"} Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.521296 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.627237 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckw8g\" (UniqueName: \"kubernetes.io/projected/9ff0f1b1-a485-45a1-8a01-e59eb360445d-kube-api-access-ckw8g\") pod \"9ff0f1b1-a485-45a1-8a01-e59eb360445d\" (UID: \"9ff0f1b1-a485-45a1-8a01-e59eb360445d\") " Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.628572 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-258661ee-6084-4970-8b6a-a82e69415274\") pod \"9ff0f1b1-a485-45a1-8a01-e59eb360445d\" (UID: \"9ff0f1b1-a485-45a1-8a01-e59eb360445d\") " Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.634964 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ff0f1b1-a485-45a1-8a01-e59eb360445d-kube-api-access-ckw8g" (OuterVolumeSpecName: "kube-api-access-ckw8g") pod "9ff0f1b1-a485-45a1-8a01-e59eb360445d" (UID: "9ff0f1b1-a485-45a1-8a01-e59eb360445d"). InnerVolumeSpecName "kube-api-access-ckw8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.652886 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-258661ee-6084-4970-8b6a-a82e69415274" (OuterVolumeSpecName: "mariadb-data") pod "9ff0f1b1-a485-45a1-8a01-e59eb360445d" (UID: "9ff0f1b1-a485-45a1-8a01-e59eb360445d"). InnerVolumeSpecName "pvc-258661ee-6084-4970-8b6a-a82e69415274". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.731282 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-258661ee-6084-4970-8b6a-a82e69415274\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-258661ee-6084-4970-8b6a-a82e69415274\") on node \"crc\" " Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.731316 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckw8g\" (UniqueName: \"kubernetes.io/projected/9ff0f1b1-a485-45a1-8a01-e59eb360445d-kube-api-access-ckw8g\") on node \"crc\" DevicePath \"\"" Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.757691 4944 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.757840 4944 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-258661ee-6084-4970-8b6a-a82e69415274" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-258661ee-6084-4970-8b6a-a82e69415274") on node "crc" Nov 24 11:28:54 crc kubenswrapper[4944]: I1124 11:28:54.832956 4944 reconciler_common.go:293] "Volume detached for volume \"pvc-258661ee-6084-4970-8b6a-a82e69415274\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-258661ee-6084-4970-8b6a-a82e69415274\") on node \"crc\" DevicePath \"\"" Nov 24 11:28:55 crc kubenswrapper[4944]: I1124 11:28:55.130961 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"9ff0f1b1-a485-45a1-8a01-e59eb360445d","Type":"ContainerDied","Data":"93f2df2de8ef452b3506018300d9e73624494ec2ac2fdcf9a6c3655efc913020"} Nov 24 11:28:55 crc kubenswrapper[4944]: I1124 11:28:55.131011 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 24 11:28:55 crc kubenswrapper[4944]: I1124 11:28:55.131027 4944 scope.go:117] "RemoveContainer" containerID="4c97f4fa04a0fb4a629e12fd1304cd09082f79eaebd148be0cdabcea44912e02" Nov 24 11:28:55 crc kubenswrapper[4944]: I1124 11:28:55.178277 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 11:28:55 crc kubenswrapper[4944]: I1124 11:28:55.187768 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Nov 24 11:28:55 crc kubenswrapper[4944]: I1124 11:28:55.921680 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 11:28:55 crc kubenswrapper[4944]: I1124 11:28:55.922221 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="9c64cab4-f4a0-4800-b401-3f3d107dfce1" containerName="adoption" containerID="cri-o://97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745" gracePeriod=30 Nov 24 11:28:56 crc kubenswrapper[4944]: I1124 11:28:56.277413 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:28:56 crc kubenswrapper[4944]: E1124 11:28:56.277724 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:28:56 crc kubenswrapper[4944]: I1124 11:28:56.294003 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ff0f1b1-a485-45a1-8a01-e59eb360445d" path="/var/lib/kubelet/pods/9ff0f1b1-a485-45a1-8a01-e59eb360445d/volumes" Nov 24 11:29:11 crc kubenswrapper[4944]: I1124 11:29:11.277309 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:29:11 crc kubenswrapper[4944]: E1124 11:29:11.277923 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:29:25 crc kubenswrapper[4944]: I1124 11:29:25.276926 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:29:25 crc kubenswrapper[4944]: E1124 11:29:25.277695 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.413653 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.441133 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.441197 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"9c64cab4-f4a0-4800-b401-3f3d107dfce1","Type":"ContainerDied","Data":"97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745"} Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.441254 4944 scope.go:117] "RemoveContainer" containerID="97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.441030 4944 generic.go:334] "Generic (PLEG): container finished" podID="9c64cab4-f4a0-4800-b401-3f3d107dfce1" containerID="97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745" exitCode=137 Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.441707 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"9c64cab4-f4a0-4800-b401-3f3d107dfce1","Type":"ContainerDied","Data":"7ee0f4d42d27fdd1c3967c411e735675a0565701b6581515c2b6a66a181e203c"} Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.470717 4944 scope.go:117] "RemoveContainer" containerID="97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745" Nov 24 11:29:26 crc kubenswrapper[4944]: E1124 11:29:26.471166 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745\": container with ID starting with 97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745 not found: ID does not exist" containerID="97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.471203 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745"} err="failed to get container status \"97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745\": rpc error: code = NotFound desc = could not find container \"97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745\": container with ID starting with 97f60615e492401a4d699b9a31b929c855ed52faf1bf3bcdc06b3b80eb9ce745 not found: ID does not exist" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.531656 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/9c64cab4-f4a0-4800-b401-3f3d107dfce1-ovn-data-cert\") pod \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.532581 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04716da7-9e6d-4a95-810d-86d54b500df3\") pod \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.532786 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gcnz\" (UniqueName: \"kubernetes.io/projected/9c64cab4-f4a0-4800-b401-3f3d107dfce1-kube-api-access-4gcnz\") pod \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\" (UID: \"9c64cab4-f4a0-4800-b401-3f3d107dfce1\") " Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.538996 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c64cab4-f4a0-4800-b401-3f3d107dfce1-kube-api-access-4gcnz" (OuterVolumeSpecName: "kube-api-access-4gcnz") pod "9c64cab4-f4a0-4800-b401-3f3d107dfce1" (UID: "9c64cab4-f4a0-4800-b401-3f3d107dfce1"). InnerVolumeSpecName "kube-api-access-4gcnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.544151 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c64cab4-f4a0-4800-b401-3f3d107dfce1-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "9c64cab4-f4a0-4800-b401-3f3d107dfce1" (UID: "9c64cab4-f4a0-4800-b401-3f3d107dfce1"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.547335 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04716da7-9e6d-4a95-810d-86d54b500df3" (OuterVolumeSpecName: "ovn-data") pod "9c64cab4-f4a0-4800-b401-3f3d107dfce1" (UID: "9c64cab4-f4a0-4800-b401-3f3d107dfce1"). InnerVolumeSpecName "pvc-04716da7-9e6d-4a95-810d-86d54b500df3". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.636660 4944 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/9c64cab4-f4a0-4800-b401-3f3d107dfce1-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.636728 4944 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-04716da7-9e6d-4a95-810d-86d54b500df3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04716da7-9e6d-4a95-810d-86d54b500df3\") on node \"crc\" " Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.636742 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gcnz\" (UniqueName: \"kubernetes.io/projected/9c64cab4-f4a0-4800-b401-3f3d107dfce1-kube-api-access-4gcnz\") on node \"crc\" DevicePath \"\"" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.660255 4944 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.660454 4944 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-04716da7-9e6d-4a95-810d-86d54b500df3" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04716da7-9e6d-4a95-810d-86d54b500df3") on node "crc" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.739246 4944 reconciler_common.go:293] "Volume detached for volume \"pvc-04716da7-9e6d-4a95-810d-86d54b500df3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-04716da7-9e6d-4a95-810d-86d54b500df3\") on node \"crc\" DevicePath \"\"" Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.781930 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 11:29:26 crc kubenswrapper[4944]: I1124 11:29:26.790883 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Nov 24 11:29:28 crc kubenswrapper[4944]: I1124 11:29:28.291744 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c64cab4-f4a0-4800-b401-3f3d107dfce1" path="/var/lib/kubelet/pods/9c64cab4-f4a0-4800-b401-3f3d107dfce1/volumes" Nov 24 11:29:39 crc kubenswrapper[4944]: I1124 11:29:39.278240 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:29:39 crc kubenswrapper[4944]: E1124 11:29:39.279032 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:29:54 crc kubenswrapper[4944]: I1124 11:29:54.277545 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:29:54 crc kubenswrapper[4944]: I1124 11:29:54.718377 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"e2aa81f7eceb76b67b33340bde645da1f586279fcd8142d56e3e89877f97e54b"} Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.155259 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4"] Nov 24 11:30:00 crc kubenswrapper[4944]: E1124 11:30:00.156219 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf413e07-7450-4eb5-8ff3-3950c9e0d44a" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156237 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf413e07-7450-4eb5-8ff3-3950c9e0d44a" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 24 11:30:00 crc kubenswrapper[4944]: E1124 11:30:00.156250 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" containerName="extract-content" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156257 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" containerName="extract-content" Nov 24 11:30:00 crc kubenswrapper[4944]: E1124 11:30:00.156278 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c64cab4-f4a0-4800-b401-3f3d107dfce1" containerName="adoption" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156354 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c64cab4-f4a0-4800-b401-3f3d107dfce1" containerName="adoption" Nov 24 11:30:00 crc kubenswrapper[4944]: E1124 11:30:00.156375 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerName="extract-utilities" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156384 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerName="extract-utilities" Nov 24 11:30:00 crc kubenswrapper[4944]: E1124 11:30:00.156403 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerName="registry-server" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156409 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerName="registry-server" Nov 24 11:30:00 crc kubenswrapper[4944]: E1124 11:30:00.156425 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ff0f1b1-a485-45a1-8a01-e59eb360445d" containerName="adoption" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156431 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ff0f1b1-a485-45a1-8a01-e59eb360445d" containerName="adoption" Nov 24 11:30:00 crc kubenswrapper[4944]: E1124 11:30:00.156446 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" containerName="registry-server" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156451 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" containerName="registry-server" Nov 24 11:30:00 crc kubenswrapper[4944]: E1124 11:30:00.156469 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" containerName="extract-utilities" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156475 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" containerName="extract-utilities" Nov 24 11:30:00 crc kubenswrapper[4944]: E1124 11:30:00.156486 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerName="extract-content" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156492 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerName="extract-content" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156743 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ff0f1b1-a485-45a1-8a01-e59eb360445d" containerName="adoption" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156763 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="2425953b-57c0-4d80-b41c-c6af84bfa1bf" containerName="registry-server" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156773 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="66850c80-f50b-413d-aec6-2dfcc9d3a633" containerName="registry-server" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156783 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf413e07-7450-4eb5-8ff3-3950c9e0d44a" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.156827 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c64cab4-f4a0-4800-b401-3f3d107dfce1" containerName="adoption" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.157847 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.160428 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.161397 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.164729 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4"] Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.272722 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a821b05-00ae-4d32-918e-451bb0aeadcb-secret-volume\") pod \"collect-profiles-29399730-tbjq4\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.272831 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n27p\" (UniqueName: \"kubernetes.io/projected/1a821b05-00ae-4d32-918e-451bb0aeadcb-kube-api-access-2n27p\") pod \"collect-profiles-29399730-tbjq4\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.273078 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a821b05-00ae-4d32-918e-451bb0aeadcb-config-volume\") pod \"collect-profiles-29399730-tbjq4\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.375293 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a821b05-00ae-4d32-918e-451bb0aeadcb-secret-volume\") pod \"collect-profiles-29399730-tbjq4\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.375368 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n27p\" (UniqueName: \"kubernetes.io/projected/1a821b05-00ae-4d32-918e-451bb0aeadcb-kube-api-access-2n27p\") pod \"collect-profiles-29399730-tbjq4\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.375424 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a821b05-00ae-4d32-918e-451bb0aeadcb-config-volume\") pod \"collect-profiles-29399730-tbjq4\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.376622 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a821b05-00ae-4d32-918e-451bb0aeadcb-config-volume\") pod \"collect-profiles-29399730-tbjq4\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.384743 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a821b05-00ae-4d32-918e-451bb0aeadcb-secret-volume\") pod \"collect-profiles-29399730-tbjq4\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.393070 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n27p\" (UniqueName: \"kubernetes.io/projected/1a821b05-00ae-4d32-918e-451bb0aeadcb-kube-api-access-2n27p\") pod \"collect-profiles-29399730-tbjq4\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.482021 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:00 crc kubenswrapper[4944]: I1124 11:30:00.978016 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4"] Nov 24 11:30:01 crc kubenswrapper[4944]: I1124 11:30:01.784728 4944 generic.go:334] "Generic (PLEG): container finished" podID="1a821b05-00ae-4d32-918e-451bb0aeadcb" containerID="40412615e739e15b64d7318f20383ef36046fab6272f4a071a4955c89d48cce6" exitCode=0 Nov 24 11:30:01 crc kubenswrapper[4944]: I1124 11:30:01.784849 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" event={"ID":"1a821b05-00ae-4d32-918e-451bb0aeadcb","Type":"ContainerDied","Data":"40412615e739e15b64d7318f20383ef36046fab6272f4a071a4955c89d48cce6"} Nov 24 11:30:01 crc kubenswrapper[4944]: I1124 11:30:01.785272 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" event={"ID":"1a821b05-00ae-4d32-918e-451bb0aeadcb","Type":"ContainerStarted","Data":"478bdbd38e1d454d0ac49efa064f9753bcb52504e1696bb6d6a00e30d8c814cb"} Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.172287 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.333952 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n27p\" (UniqueName: \"kubernetes.io/projected/1a821b05-00ae-4d32-918e-451bb0aeadcb-kube-api-access-2n27p\") pod \"1a821b05-00ae-4d32-918e-451bb0aeadcb\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.334119 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a821b05-00ae-4d32-918e-451bb0aeadcb-config-volume\") pod \"1a821b05-00ae-4d32-918e-451bb0aeadcb\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.334237 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a821b05-00ae-4d32-918e-451bb0aeadcb-secret-volume\") pod \"1a821b05-00ae-4d32-918e-451bb0aeadcb\" (UID: \"1a821b05-00ae-4d32-918e-451bb0aeadcb\") " Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.335561 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a821b05-00ae-4d32-918e-451bb0aeadcb-config-volume" (OuterVolumeSpecName: "config-volume") pod "1a821b05-00ae-4d32-918e-451bb0aeadcb" (UID: "1a821b05-00ae-4d32-918e-451bb0aeadcb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.342789 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a821b05-00ae-4d32-918e-451bb0aeadcb-kube-api-access-2n27p" (OuterVolumeSpecName: "kube-api-access-2n27p") pod "1a821b05-00ae-4d32-918e-451bb0aeadcb" (UID: "1a821b05-00ae-4d32-918e-451bb0aeadcb"). InnerVolumeSpecName "kube-api-access-2n27p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.353661 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a821b05-00ae-4d32-918e-451bb0aeadcb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1a821b05-00ae-4d32-918e-451bb0aeadcb" (UID: "1a821b05-00ae-4d32-918e-451bb0aeadcb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.436997 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n27p\" (UniqueName: \"kubernetes.io/projected/1a821b05-00ae-4d32-918e-451bb0aeadcb-kube-api-access-2n27p\") on node \"crc\" DevicePath \"\"" Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.437034 4944 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1a821b05-00ae-4d32-918e-451bb0aeadcb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.437046 4944 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1a821b05-00ae-4d32-918e-451bb0aeadcb-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.822932 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" event={"ID":"1a821b05-00ae-4d32-918e-451bb0aeadcb","Type":"ContainerDied","Data":"478bdbd38e1d454d0ac49efa064f9753bcb52504e1696bb6d6a00e30d8c814cb"} Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.823300 4944 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="478bdbd38e1d454d0ac49efa064f9753bcb52504e1696bb6d6a00e30d8c814cb" Nov 24 11:30:03 crc kubenswrapper[4944]: I1124 11:30:03.822983 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399730-tbjq4" Nov 24 11:30:04 crc kubenswrapper[4944]: I1124 11:30:04.261571 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5"] Nov 24 11:30:04 crc kubenswrapper[4944]: I1124 11:30:04.271612 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399685-8djd5"] Nov 24 11:30:04 crc kubenswrapper[4944]: I1124 11:30:04.289439 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0" path="/var/lib/kubelet/pods/b2ecd1e5-f734-4c80-a1e7-bcf98e806ea0/volumes" Nov 24 11:30:26 crc kubenswrapper[4944]: I1124 11:30:26.818814 4944 trace.go:236] Trace[986723649]: "Calculate volume metrics of mysql-db for pod openstack/openstack-galera-0" (24-Nov-2025 11:30:25.815) (total time: 1003ms): Nov 24 11:30:26 crc kubenswrapper[4944]: Trace[986723649]: [1.003067985s] [1.003067985s] END Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.470905 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gqft2/must-gather-6kf28"] Nov 24 11:30:32 crc kubenswrapper[4944]: E1124 11:30:32.472042 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a821b05-00ae-4d32-918e-451bb0aeadcb" containerName="collect-profiles" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.472075 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a821b05-00ae-4d32-918e-451bb0aeadcb" containerName="collect-profiles" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.472338 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a821b05-00ae-4d32-918e-451bb0aeadcb" containerName="collect-profiles" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.480085 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/must-gather-6kf28" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.482536 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gqft2"/"kube-root-ca.crt" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.492754 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gqft2/must-gather-6kf28"] Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.500806 4944 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gqft2"/"openshift-service-ca.crt" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.587504 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv569\" (UniqueName: \"kubernetes.io/projected/e39efbba-63ef-4598-98ca-e88d7ca05ce7-kube-api-access-fv569\") pod \"must-gather-6kf28\" (UID: \"e39efbba-63ef-4598-98ca-e88d7ca05ce7\") " pod="openshift-must-gather-gqft2/must-gather-6kf28" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.587852 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e39efbba-63ef-4598-98ca-e88d7ca05ce7-must-gather-output\") pod \"must-gather-6kf28\" (UID: \"e39efbba-63ef-4598-98ca-e88d7ca05ce7\") " pod="openshift-must-gather-gqft2/must-gather-6kf28" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.690213 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv569\" (UniqueName: \"kubernetes.io/projected/e39efbba-63ef-4598-98ca-e88d7ca05ce7-kube-api-access-fv569\") pod \"must-gather-6kf28\" (UID: \"e39efbba-63ef-4598-98ca-e88d7ca05ce7\") " pod="openshift-must-gather-gqft2/must-gather-6kf28" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.690402 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e39efbba-63ef-4598-98ca-e88d7ca05ce7-must-gather-output\") pod \"must-gather-6kf28\" (UID: \"e39efbba-63ef-4598-98ca-e88d7ca05ce7\") " pod="openshift-must-gather-gqft2/must-gather-6kf28" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.690826 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e39efbba-63ef-4598-98ca-e88d7ca05ce7-must-gather-output\") pod \"must-gather-6kf28\" (UID: \"e39efbba-63ef-4598-98ca-e88d7ca05ce7\") " pod="openshift-must-gather-gqft2/must-gather-6kf28" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.709844 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv569\" (UniqueName: \"kubernetes.io/projected/e39efbba-63ef-4598-98ca-e88d7ca05ce7-kube-api-access-fv569\") pod \"must-gather-6kf28\" (UID: \"e39efbba-63ef-4598-98ca-e88d7ca05ce7\") " pod="openshift-must-gather-gqft2/must-gather-6kf28" Nov 24 11:30:32 crc kubenswrapper[4944]: I1124 11:30:32.806074 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/must-gather-6kf28" Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.322882 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gqft2/must-gather-6kf28"] Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.332936 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.539995 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dzlsb"] Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.543022 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.557007 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzlsb"] Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.659343 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-utilities\") pod \"redhat-marketplace-dzlsb\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.659459 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqrl7\" (UniqueName: \"kubernetes.io/projected/e1cb84d1-5aef-41b1-9891-91d807cae64f-kube-api-access-sqrl7\") pod \"redhat-marketplace-dzlsb\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.659593 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-catalog-content\") pod \"redhat-marketplace-dzlsb\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.761896 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-utilities\") pod \"redhat-marketplace-dzlsb\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.761996 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqrl7\" (UniqueName: \"kubernetes.io/projected/e1cb84d1-5aef-41b1-9891-91d807cae64f-kube-api-access-sqrl7\") pod \"redhat-marketplace-dzlsb\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.762133 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-catalog-content\") pod \"redhat-marketplace-dzlsb\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.762668 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-catalog-content\") pod \"redhat-marketplace-dzlsb\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.762778 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-utilities\") pod \"redhat-marketplace-dzlsb\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.790901 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqrl7\" (UniqueName: \"kubernetes.io/projected/e1cb84d1-5aef-41b1-9891-91d807cae64f-kube-api-access-sqrl7\") pod \"redhat-marketplace-dzlsb\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:33 crc kubenswrapper[4944]: I1124 11:30:33.873303 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:34 crc kubenswrapper[4944]: I1124 11:30:34.152758 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqft2/must-gather-6kf28" event={"ID":"e39efbba-63ef-4598-98ca-e88d7ca05ce7","Type":"ContainerStarted","Data":"ac26f4e77700e1c21e72458d48429b7d14dbddab1567bf8887e211a56568f380"} Nov 24 11:30:34 crc kubenswrapper[4944]: I1124 11:30:34.436850 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzlsb"] Nov 24 11:30:35 crc kubenswrapper[4944]: I1124 11:30:35.173549 4944 generic.go:334] "Generic (PLEG): container finished" podID="e1cb84d1-5aef-41b1-9891-91d807cae64f" containerID="725fd33157fa6e4e2c3d5467c42870397893847288f5b7c826d1712d494592db" exitCode=0 Nov 24 11:30:35 crc kubenswrapper[4944]: I1124 11:30:35.173858 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzlsb" event={"ID":"e1cb84d1-5aef-41b1-9891-91d807cae64f","Type":"ContainerDied","Data":"725fd33157fa6e4e2c3d5467c42870397893847288f5b7c826d1712d494592db"} Nov 24 11:30:35 crc kubenswrapper[4944]: I1124 11:30:35.173884 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzlsb" event={"ID":"e1cb84d1-5aef-41b1-9891-91d807cae64f","Type":"ContainerStarted","Data":"d6048a97f46f80c2674169e42a641175d7a17c5307bf08b74b3e1d017b5c67e9"} Nov 24 11:30:37 crc kubenswrapper[4944]: I1124 11:30:37.202634 4944 generic.go:334] "Generic (PLEG): container finished" podID="e1cb84d1-5aef-41b1-9891-91d807cae64f" containerID="e07d578d94cbc15fac9bb69c5785b83eaf03e131c1b19102573286bac1ac25b0" exitCode=0 Nov 24 11:30:37 crc kubenswrapper[4944]: I1124 11:30:37.202701 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzlsb" event={"ID":"e1cb84d1-5aef-41b1-9891-91d807cae64f","Type":"ContainerDied","Data":"e07d578d94cbc15fac9bb69c5785b83eaf03e131c1b19102573286bac1ac25b0"} Nov 24 11:30:37 crc kubenswrapper[4944]: I1124 11:30:37.393553 4944 scope.go:117] "RemoveContainer" containerID="90075532c080473cc712253ebd5c712e14bc3c81827c9a1b645898014a12a778" Nov 24 11:30:39 crc kubenswrapper[4944]: I1124 11:30:39.226979 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzlsb" event={"ID":"e1cb84d1-5aef-41b1-9891-91d807cae64f","Type":"ContainerStarted","Data":"9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde"} Nov 24 11:30:39 crc kubenswrapper[4944]: I1124 11:30:39.229764 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqft2/must-gather-6kf28" event={"ID":"e39efbba-63ef-4598-98ca-e88d7ca05ce7","Type":"ContainerStarted","Data":"659ae4b751b5491017c6a35b523375044a9951da98f218bc5ba44933ddeaff94"} Nov 24 11:30:39 crc kubenswrapper[4944]: I1124 11:30:39.229814 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqft2/must-gather-6kf28" event={"ID":"e39efbba-63ef-4598-98ca-e88d7ca05ce7","Type":"ContainerStarted","Data":"ca05f81c5ee2cf26da8b2faca63e17d62ed1a742f3b14e143a906716bafc2fda"} Nov 24 11:30:39 crc kubenswrapper[4944]: I1124 11:30:39.249477 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dzlsb" podStartSLOduration=2.746239918 podStartE2EDuration="6.249458651s" podCreationTimestamp="2025-11-24 11:30:33 +0000 UTC" firstStartedPulling="2025-11-24 11:30:35.181505211 +0000 UTC m=+9495.715945673" lastFinishedPulling="2025-11-24 11:30:38.684723944 +0000 UTC m=+9499.219164406" observedRunningTime="2025-11-24 11:30:39.247734517 +0000 UTC m=+9499.782175029" watchObservedRunningTime="2025-11-24 11:30:39.249458651 +0000 UTC m=+9499.783899113" Nov 24 11:30:39 crc kubenswrapper[4944]: I1124 11:30:39.265854 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gqft2/must-gather-6kf28" podStartSLOduration=2.290143128 podStartE2EDuration="7.265835227s" podCreationTimestamp="2025-11-24 11:30:32 +0000 UTC" firstStartedPulling="2025-11-24 11:30:33.332896198 +0000 UTC m=+9493.867336670" lastFinishedPulling="2025-11-24 11:30:38.308588307 +0000 UTC m=+9498.843028769" observedRunningTime="2025-11-24 11:30:39.262459231 +0000 UTC m=+9499.796899693" watchObservedRunningTime="2025-11-24 11:30:39.265835227 +0000 UTC m=+9499.800275689" Nov 24 11:30:42 crc kubenswrapper[4944]: I1124 11:30:42.927707 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gqft2/crc-debug-n98nj"] Nov 24 11:30:42 crc kubenswrapper[4944]: I1124 11:30:42.931544 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/crc-debug-n98nj" Nov 24 11:30:42 crc kubenswrapper[4944]: I1124 11:30:42.939476 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gqft2"/"default-dockercfg-c997n" Nov 24 11:30:43 crc kubenswrapper[4944]: I1124 11:30:43.096416 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlxpw\" (UniqueName: \"kubernetes.io/projected/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-kube-api-access-dlxpw\") pod \"crc-debug-n98nj\" (UID: \"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02\") " pod="openshift-must-gather-gqft2/crc-debug-n98nj" Nov 24 11:30:43 crc kubenswrapper[4944]: I1124 11:30:43.096925 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-host\") pod \"crc-debug-n98nj\" (UID: \"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02\") " pod="openshift-must-gather-gqft2/crc-debug-n98nj" Nov 24 11:30:43 crc kubenswrapper[4944]: I1124 11:30:43.199543 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlxpw\" (UniqueName: \"kubernetes.io/projected/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-kube-api-access-dlxpw\") pod \"crc-debug-n98nj\" (UID: \"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02\") " pod="openshift-must-gather-gqft2/crc-debug-n98nj" Nov 24 11:30:43 crc kubenswrapper[4944]: I1124 11:30:43.199617 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-host\") pod \"crc-debug-n98nj\" (UID: \"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02\") " pod="openshift-must-gather-gqft2/crc-debug-n98nj" Nov 24 11:30:43 crc kubenswrapper[4944]: I1124 11:30:43.199782 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-host\") pod \"crc-debug-n98nj\" (UID: \"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02\") " pod="openshift-must-gather-gqft2/crc-debug-n98nj" Nov 24 11:30:43 crc kubenswrapper[4944]: I1124 11:30:43.225780 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlxpw\" (UniqueName: \"kubernetes.io/projected/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-kube-api-access-dlxpw\") pod \"crc-debug-n98nj\" (UID: \"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02\") " pod="openshift-must-gather-gqft2/crc-debug-n98nj" Nov 24 11:30:43 crc kubenswrapper[4944]: I1124 11:30:43.263433 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/crc-debug-n98nj" Nov 24 11:30:43 crc kubenswrapper[4944]: W1124 11:30:43.310359 4944 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36ba8d91_21cc_4eac_9875_5d1c3ba4ee02.slice/crio-359bde021a8b5a7d8b4d557c1122096cf96ecf6aa5fcc5925e4a54220a7ca37e WatchSource:0}: Error finding container 359bde021a8b5a7d8b4d557c1122096cf96ecf6aa5fcc5925e4a54220a7ca37e: Status 404 returned error can't find the container with id 359bde021a8b5a7d8b4d557c1122096cf96ecf6aa5fcc5925e4a54220a7ca37e Nov 24 11:30:43 crc kubenswrapper[4944]: I1124 11:30:43.873525 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:43 crc kubenswrapper[4944]: I1124 11:30:43.873910 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:43 crc kubenswrapper[4944]: I1124 11:30:43.928173 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:44 crc kubenswrapper[4944]: I1124 11:30:44.309031 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqft2/crc-debug-n98nj" event={"ID":"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02","Type":"ContainerStarted","Data":"359bde021a8b5a7d8b4d557c1122096cf96ecf6aa5fcc5925e4a54220a7ca37e"} Nov 24 11:30:44 crc kubenswrapper[4944]: I1124 11:30:44.365076 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:44 crc kubenswrapper[4944]: I1124 11:30:44.420289 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzlsb"] Nov 24 11:30:46 crc kubenswrapper[4944]: I1124 11:30:46.333191 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dzlsb" podUID="e1cb84d1-5aef-41b1-9891-91d807cae64f" containerName="registry-server" containerID="cri-o://9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde" gracePeriod=2 Nov 24 11:30:46 crc kubenswrapper[4944]: I1124 11:30:46.925012 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.097184 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqrl7\" (UniqueName: \"kubernetes.io/projected/e1cb84d1-5aef-41b1-9891-91d807cae64f-kube-api-access-sqrl7\") pod \"e1cb84d1-5aef-41b1-9891-91d807cae64f\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.097368 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-catalog-content\") pod \"e1cb84d1-5aef-41b1-9891-91d807cae64f\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.097445 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-utilities\") pod \"e1cb84d1-5aef-41b1-9891-91d807cae64f\" (UID: \"e1cb84d1-5aef-41b1-9891-91d807cae64f\") " Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.098540 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-utilities" (OuterVolumeSpecName: "utilities") pod "e1cb84d1-5aef-41b1-9891-91d807cae64f" (UID: "e1cb84d1-5aef-41b1-9891-91d807cae64f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.116504 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1cb84d1-5aef-41b1-9891-91d807cae64f-kube-api-access-sqrl7" (OuterVolumeSpecName: "kube-api-access-sqrl7") pod "e1cb84d1-5aef-41b1-9891-91d807cae64f" (UID: "e1cb84d1-5aef-41b1-9891-91d807cae64f"). InnerVolumeSpecName "kube-api-access-sqrl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.119551 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1cb84d1-5aef-41b1-9891-91d807cae64f" (UID: "e1cb84d1-5aef-41b1-9891-91d807cae64f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.200001 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.200319 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1cb84d1-5aef-41b1-9891-91d807cae64f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.200406 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqrl7\" (UniqueName: \"kubernetes.io/projected/e1cb84d1-5aef-41b1-9891-91d807cae64f-kube-api-access-sqrl7\") on node \"crc\" DevicePath \"\"" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.348452 4944 generic.go:334] "Generic (PLEG): container finished" podID="e1cb84d1-5aef-41b1-9891-91d807cae64f" containerID="9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde" exitCode=0 Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.348505 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzlsb" event={"ID":"e1cb84d1-5aef-41b1-9891-91d807cae64f","Type":"ContainerDied","Data":"9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde"} Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.348542 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzlsb" event={"ID":"e1cb84d1-5aef-41b1-9891-91d807cae64f","Type":"ContainerDied","Data":"d6048a97f46f80c2674169e42a641175d7a17c5307bf08b74b3e1d017b5c67e9"} Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.348564 4944 scope.go:117] "RemoveContainer" containerID="9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.348723 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzlsb" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.391093 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzlsb"] Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.397581 4944 scope.go:117] "RemoveContainer" containerID="e07d578d94cbc15fac9bb69c5785b83eaf03e131c1b19102573286bac1ac25b0" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.401187 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzlsb"] Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.424722 4944 scope.go:117] "RemoveContainer" containerID="725fd33157fa6e4e2c3d5467c42870397893847288f5b7c826d1712d494592db" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.501692 4944 scope.go:117] "RemoveContainer" containerID="9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde" Nov 24 11:30:47 crc kubenswrapper[4944]: E1124 11:30:47.502225 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde\": container with ID starting with 9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde not found: ID does not exist" containerID="9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.502268 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde"} err="failed to get container status \"9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde\": rpc error: code = NotFound desc = could not find container \"9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde\": container with ID starting with 9ee1c2a92d4f29456f44535a255b5d42a20042ce269e792f4bef1e1d6b5e4cde not found: ID does not exist" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.502297 4944 scope.go:117] "RemoveContainer" containerID="e07d578d94cbc15fac9bb69c5785b83eaf03e131c1b19102573286bac1ac25b0" Nov 24 11:30:47 crc kubenswrapper[4944]: E1124 11:30:47.502783 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e07d578d94cbc15fac9bb69c5785b83eaf03e131c1b19102573286bac1ac25b0\": container with ID starting with e07d578d94cbc15fac9bb69c5785b83eaf03e131c1b19102573286bac1ac25b0 not found: ID does not exist" containerID="e07d578d94cbc15fac9bb69c5785b83eaf03e131c1b19102573286bac1ac25b0" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.502812 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e07d578d94cbc15fac9bb69c5785b83eaf03e131c1b19102573286bac1ac25b0"} err="failed to get container status \"e07d578d94cbc15fac9bb69c5785b83eaf03e131c1b19102573286bac1ac25b0\": rpc error: code = NotFound desc = could not find container \"e07d578d94cbc15fac9bb69c5785b83eaf03e131c1b19102573286bac1ac25b0\": container with ID starting with e07d578d94cbc15fac9bb69c5785b83eaf03e131c1b19102573286bac1ac25b0 not found: ID does not exist" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.502832 4944 scope.go:117] "RemoveContainer" containerID="725fd33157fa6e4e2c3d5467c42870397893847288f5b7c826d1712d494592db" Nov 24 11:30:47 crc kubenswrapper[4944]: E1124 11:30:47.503022 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"725fd33157fa6e4e2c3d5467c42870397893847288f5b7c826d1712d494592db\": container with ID starting with 725fd33157fa6e4e2c3d5467c42870397893847288f5b7c826d1712d494592db not found: ID does not exist" containerID="725fd33157fa6e4e2c3d5467c42870397893847288f5b7c826d1712d494592db" Nov 24 11:30:47 crc kubenswrapper[4944]: I1124 11:30:47.503040 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725fd33157fa6e4e2c3d5467c42870397893847288f5b7c826d1712d494592db"} err="failed to get container status \"725fd33157fa6e4e2c3d5467c42870397893847288f5b7c826d1712d494592db\": rpc error: code = NotFound desc = could not find container \"725fd33157fa6e4e2c3d5467c42870397893847288f5b7c826d1712d494592db\": container with ID starting with 725fd33157fa6e4e2c3d5467c42870397893847288f5b7c826d1712d494592db not found: ID does not exist" Nov 24 11:30:48 crc kubenswrapper[4944]: I1124 11:30:48.290597 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1cb84d1-5aef-41b1-9891-91d807cae64f" path="/var/lib/kubelet/pods/e1cb84d1-5aef-41b1-9891-91d807cae64f/volumes" Nov 24 11:30:56 crc kubenswrapper[4944]: I1124 11:30:56.529184 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqft2/crc-debug-n98nj" event={"ID":"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02","Type":"ContainerStarted","Data":"298a932688bcd993c6d754bd62ab8570db84588a29beff1cb1f3700ca228cd40"} Nov 24 11:30:56 crc kubenswrapper[4944]: I1124 11:30:56.560109 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gqft2/crc-debug-n98nj" podStartSLOduration=1.712355852 podStartE2EDuration="14.560089334s" podCreationTimestamp="2025-11-24 11:30:42 +0000 UTC" firstStartedPulling="2025-11-24 11:30:43.313141006 +0000 UTC m=+9503.847581468" lastFinishedPulling="2025-11-24 11:30:56.160874488 +0000 UTC m=+9516.695314950" observedRunningTime="2025-11-24 11:30:56.550193191 +0000 UTC m=+9517.084633653" watchObservedRunningTime="2025-11-24 11:30:56.560089334 +0000 UTC m=+9517.094529796" Nov 24 11:31:18 crc kubenswrapper[4944]: I1124 11:31:18.747552 4944 generic.go:334] "Generic (PLEG): container finished" podID="36ba8d91-21cc-4eac-9875-5d1c3ba4ee02" containerID="298a932688bcd993c6d754bd62ab8570db84588a29beff1cb1f3700ca228cd40" exitCode=0 Nov 24 11:31:18 crc kubenswrapper[4944]: I1124 11:31:18.747630 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqft2/crc-debug-n98nj" event={"ID":"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02","Type":"ContainerDied","Data":"298a932688bcd993c6d754bd62ab8570db84588a29beff1cb1f3700ca228cd40"} Nov 24 11:31:19 crc kubenswrapper[4944]: I1124 11:31:19.891623 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/crc-debug-n98nj" Nov 24 11:31:19 crc kubenswrapper[4944]: I1124 11:31:19.930153 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gqft2/crc-debug-n98nj"] Nov 24 11:31:19 crc kubenswrapper[4944]: I1124 11:31:19.946086 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gqft2/crc-debug-n98nj"] Nov 24 11:31:20 crc kubenswrapper[4944]: I1124 11:31:20.049655 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-host\") pod \"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02\" (UID: \"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02\") " Nov 24 11:31:20 crc kubenswrapper[4944]: I1124 11:31:20.049789 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-host" (OuterVolumeSpecName: "host") pod "36ba8d91-21cc-4eac-9875-5d1c3ba4ee02" (UID: "36ba8d91-21cc-4eac-9875-5d1c3ba4ee02"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 11:31:20 crc kubenswrapper[4944]: I1124 11:31:20.049888 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlxpw\" (UniqueName: \"kubernetes.io/projected/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-kube-api-access-dlxpw\") pod \"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02\" (UID: \"36ba8d91-21cc-4eac-9875-5d1c3ba4ee02\") " Nov 24 11:31:20 crc kubenswrapper[4944]: I1124 11:31:20.050709 4944 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-host\") on node \"crc\" DevicePath \"\"" Nov 24 11:31:20 crc kubenswrapper[4944]: I1124 11:31:20.067862 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-kube-api-access-dlxpw" (OuterVolumeSpecName: "kube-api-access-dlxpw") pod "36ba8d91-21cc-4eac-9875-5d1c3ba4ee02" (UID: "36ba8d91-21cc-4eac-9875-5d1c3ba4ee02"). InnerVolumeSpecName "kube-api-access-dlxpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:31:20 crc kubenswrapper[4944]: I1124 11:31:20.153111 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlxpw\" (UniqueName: \"kubernetes.io/projected/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02-kube-api-access-dlxpw\") on node \"crc\" DevicePath \"\"" Nov 24 11:31:20 crc kubenswrapper[4944]: I1124 11:31:20.293954 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36ba8d91-21cc-4eac-9875-5d1c3ba4ee02" path="/var/lib/kubelet/pods/36ba8d91-21cc-4eac-9875-5d1c3ba4ee02/volumes" Nov 24 11:31:20 crc kubenswrapper[4944]: I1124 11:31:20.765096 4944 scope.go:117] "RemoveContainer" containerID="298a932688bcd993c6d754bd62ab8570db84588a29beff1cb1f3700ca228cd40" Nov 24 11:31:20 crc kubenswrapper[4944]: I1124 11:31:20.765424 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/crc-debug-n98nj" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.167561 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gqft2/crc-debug-jjkml"] Nov 24 11:31:21 crc kubenswrapper[4944]: E1124 11:31:21.168362 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36ba8d91-21cc-4eac-9875-5d1c3ba4ee02" containerName="container-00" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.168374 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="36ba8d91-21cc-4eac-9875-5d1c3ba4ee02" containerName="container-00" Nov 24 11:31:21 crc kubenswrapper[4944]: E1124 11:31:21.168398 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1cb84d1-5aef-41b1-9891-91d807cae64f" containerName="registry-server" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.168404 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1cb84d1-5aef-41b1-9891-91d807cae64f" containerName="registry-server" Nov 24 11:31:21 crc kubenswrapper[4944]: E1124 11:31:21.168418 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1cb84d1-5aef-41b1-9891-91d807cae64f" containerName="extract-utilities" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.168424 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1cb84d1-5aef-41b1-9891-91d807cae64f" containerName="extract-utilities" Nov 24 11:31:21 crc kubenswrapper[4944]: E1124 11:31:21.168438 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1cb84d1-5aef-41b1-9891-91d807cae64f" containerName="extract-content" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.168443 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1cb84d1-5aef-41b1-9891-91d807cae64f" containerName="extract-content" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.168636 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1cb84d1-5aef-41b1-9891-91d807cae64f" containerName="registry-server" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.168647 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="36ba8d91-21cc-4eac-9875-5d1c3ba4ee02" containerName="container-00" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.169347 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/crc-debug-jjkml" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.171649 4944 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gqft2"/"default-dockercfg-c997n" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.277024 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqt8x\" (UniqueName: \"kubernetes.io/projected/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-kube-api-access-sqt8x\") pod \"crc-debug-jjkml\" (UID: \"92f104b8-fd52-4b32-8d52-a198a9bc0ef2\") " pod="openshift-must-gather-gqft2/crc-debug-jjkml" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.277121 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-host\") pod \"crc-debug-jjkml\" (UID: \"92f104b8-fd52-4b32-8d52-a198a9bc0ef2\") " pod="openshift-must-gather-gqft2/crc-debug-jjkml" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.378546 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqt8x\" (UniqueName: \"kubernetes.io/projected/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-kube-api-access-sqt8x\") pod \"crc-debug-jjkml\" (UID: \"92f104b8-fd52-4b32-8d52-a198a9bc0ef2\") " pod="openshift-must-gather-gqft2/crc-debug-jjkml" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.379205 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-host\") pod \"crc-debug-jjkml\" (UID: \"92f104b8-fd52-4b32-8d52-a198a9bc0ef2\") " pod="openshift-must-gather-gqft2/crc-debug-jjkml" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.379490 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-host\") pod \"crc-debug-jjkml\" (UID: \"92f104b8-fd52-4b32-8d52-a198a9bc0ef2\") " pod="openshift-must-gather-gqft2/crc-debug-jjkml" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.419391 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqt8x\" (UniqueName: \"kubernetes.io/projected/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-kube-api-access-sqt8x\") pod \"crc-debug-jjkml\" (UID: \"92f104b8-fd52-4b32-8d52-a198a9bc0ef2\") " pod="openshift-must-gather-gqft2/crc-debug-jjkml" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.491229 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/crc-debug-jjkml" Nov 24 11:31:21 crc kubenswrapper[4944]: I1124 11:31:21.776875 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqft2/crc-debug-jjkml" event={"ID":"92f104b8-fd52-4b32-8d52-a198a9bc0ef2","Type":"ContainerStarted","Data":"b5a85b2a8c4c5289b0a23b789893490a2abafefc7a2577afed5ead5628d6c531"} Nov 24 11:31:22 crc kubenswrapper[4944]: I1124 11:31:22.790431 4944 generic.go:334] "Generic (PLEG): container finished" podID="92f104b8-fd52-4b32-8d52-a198a9bc0ef2" containerID="a857cbff3e71eb81e93860afefb8cd0b32054b78c5137f2ce7dfb8a055883cce" exitCode=1 Nov 24 11:31:22 crc kubenswrapper[4944]: I1124 11:31:22.790482 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqft2/crc-debug-jjkml" event={"ID":"92f104b8-fd52-4b32-8d52-a198a9bc0ef2","Type":"ContainerDied","Data":"a857cbff3e71eb81e93860afefb8cd0b32054b78c5137f2ce7dfb8a055883cce"} Nov 24 11:31:22 crc kubenswrapper[4944]: I1124 11:31:22.834503 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gqft2/crc-debug-jjkml"] Nov 24 11:31:22 crc kubenswrapper[4944]: I1124 11:31:22.842687 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gqft2/crc-debug-jjkml"] Nov 24 11:31:24 crc kubenswrapper[4944]: I1124 11:31:24.414341 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/crc-debug-jjkml" Nov 24 11:31:24 crc kubenswrapper[4944]: I1124 11:31:24.554962 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqt8x\" (UniqueName: \"kubernetes.io/projected/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-kube-api-access-sqt8x\") pod \"92f104b8-fd52-4b32-8d52-a198a9bc0ef2\" (UID: \"92f104b8-fd52-4b32-8d52-a198a9bc0ef2\") " Nov 24 11:31:24 crc kubenswrapper[4944]: I1124 11:31:24.555234 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-host\") pod \"92f104b8-fd52-4b32-8d52-a198a9bc0ef2\" (UID: \"92f104b8-fd52-4b32-8d52-a198a9bc0ef2\") " Nov 24 11:31:24 crc kubenswrapper[4944]: I1124 11:31:24.555350 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-host" (OuterVolumeSpecName: "host") pod "92f104b8-fd52-4b32-8d52-a198a9bc0ef2" (UID: "92f104b8-fd52-4b32-8d52-a198a9bc0ef2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 11:31:24 crc kubenswrapper[4944]: I1124 11:31:24.555819 4944 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-host\") on node \"crc\" DevicePath \"\"" Nov 24 11:31:24 crc kubenswrapper[4944]: I1124 11:31:24.568295 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-kube-api-access-sqt8x" (OuterVolumeSpecName: "kube-api-access-sqt8x") pod "92f104b8-fd52-4b32-8d52-a198a9bc0ef2" (UID: "92f104b8-fd52-4b32-8d52-a198a9bc0ef2"). InnerVolumeSpecName "kube-api-access-sqt8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:31:24 crc kubenswrapper[4944]: I1124 11:31:24.657474 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqt8x\" (UniqueName: \"kubernetes.io/projected/92f104b8-fd52-4b32-8d52-a198a9bc0ef2-kube-api-access-sqt8x\") on node \"crc\" DevicePath \"\"" Nov 24 11:31:24 crc kubenswrapper[4944]: I1124 11:31:24.811383 4944 scope.go:117] "RemoveContainer" containerID="a857cbff3e71eb81e93860afefb8cd0b32054b78c5137f2ce7dfb8a055883cce" Nov 24 11:31:24 crc kubenswrapper[4944]: I1124 11:31:24.811449 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/crc-debug-jjkml" Nov 24 11:31:26 crc kubenswrapper[4944]: I1124 11:31:26.317863 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92f104b8-fd52-4b32-8d52-a198a9bc0ef2" path="/var/lib/kubelet/pods/92f104b8-fd52-4b32-8d52-a198a9bc0ef2/volumes" Nov 24 11:32:23 crc kubenswrapper[4944]: I1124 11:32:23.548852 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:32:23 crc kubenswrapper[4944]: I1124 11:32:23.550481 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:32:53 crc kubenswrapper[4944]: I1124 11:32:53.548617 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:32:53 crc kubenswrapper[4944]: I1124 11:32:53.549721 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:33:04 crc kubenswrapper[4944]: I1124 11:33:04.899202 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zng5q"] Nov 24 11:33:04 crc kubenswrapper[4944]: E1124 11:33:04.900411 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f104b8-fd52-4b32-8d52-a198a9bc0ef2" containerName="container-00" Nov 24 11:33:04 crc kubenswrapper[4944]: I1124 11:33:04.900434 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f104b8-fd52-4b32-8d52-a198a9bc0ef2" containerName="container-00" Nov 24 11:33:04 crc kubenswrapper[4944]: I1124 11:33:04.900764 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="92f104b8-fd52-4b32-8d52-a198a9bc0ef2" containerName="container-00" Nov 24 11:33:04 crc kubenswrapper[4944]: I1124 11:33:04.903193 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:04 crc kubenswrapper[4944]: I1124 11:33:04.930215 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zng5q"] Nov 24 11:33:05 crc kubenswrapper[4944]: I1124 11:33:05.040914 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-utilities\") pod \"certified-operators-zng5q\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:05 crc kubenswrapper[4944]: I1124 11:33:05.041403 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmtw7\" (UniqueName: \"kubernetes.io/projected/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-kube-api-access-vmtw7\") pod \"certified-operators-zng5q\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:05 crc kubenswrapper[4944]: I1124 11:33:05.041516 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-catalog-content\") pod \"certified-operators-zng5q\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:05 crc kubenswrapper[4944]: I1124 11:33:05.144313 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-utilities\") pod \"certified-operators-zng5q\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:05 crc kubenswrapper[4944]: I1124 11:33:05.144461 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmtw7\" (UniqueName: \"kubernetes.io/projected/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-kube-api-access-vmtw7\") pod \"certified-operators-zng5q\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:05 crc kubenswrapper[4944]: I1124 11:33:05.144493 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-catalog-content\") pod \"certified-operators-zng5q\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:05 crc kubenswrapper[4944]: I1124 11:33:05.144998 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-catalog-content\") pod \"certified-operators-zng5q\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:05 crc kubenswrapper[4944]: I1124 11:33:05.145181 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-utilities\") pod \"certified-operators-zng5q\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:05 crc kubenswrapper[4944]: I1124 11:33:05.183343 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmtw7\" (UniqueName: \"kubernetes.io/projected/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-kube-api-access-vmtw7\") pod \"certified-operators-zng5q\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:05 crc kubenswrapper[4944]: I1124 11:33:05.226131 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:05 crc kubenswrapper[4944]: I1124 11:33:05.931274 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zng5q"] Nov 24 11:33:06 crc kubenswrapper[4944]: I1124 11:33:06.936558 4944 generic.go:334] "Generic (PLEG): container finished" podID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" containerID="301cfb1f73d5fc88721b5bc804fae8ac268cb43daab6213aab9530d7c8e75c96" exitCode=0 Nov 24 11:33:06 crc kubenswrapper[4944]: I1124 11:33:06.936599 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zng5q" event={"ID":"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e","Type":"ContainerDied","Data":"301cfb1f73d5fc88721b5bc804fae8ac268cb43daab6213aab9530d7c8e75c96"} Nov 24 11:33:06 crc kubenswrapper[4944]: I1124 11:33:06.936840 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zng5q" event={"ID":"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e","Type":"ContainerStarted","Data":"bacc4ee91526c501c7c848411df73b578693e9fe0b38df35a5ef1cb7758ff7c2"} Nov 24 11:33:07 crc kubenswrapper[4944]: I1124 11:33:07.948995 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zng5q" event={"ID":"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e","Type":"ContainerStarted","Data":"5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d"} Nov 24 11:33:09 crc kubenswrapper[4944]: I1124 11:33:09.970922 4944 generic.go:334] "Generic (PLEG): container finished" podID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" containerID="5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d" exitCode=0 Nov 24 11:33:09 crc kubenswrapper[4944]: I1124 11:33:09.971036 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zng5q" event={"ID":"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e","Type":"ContainerDied","Data":"5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d"} Nov 24 11:33:10 crc kubenswrapper[4944]: I1124 11:33:10.982967 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zng5q" event={"ID":"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e","Type":"ContainerStarted","Data":"a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182"} Nov 24 11:33:11 crc kubenswrapper[4944]: I1124 11:33:11.015731 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zng5q" podStartSLOduration=3.257869155 podStartE2EDuration="7.015693971s" podCreationTimestamp="2025-11-24 11:33:04 +0000 UTC" firstStartedPulling="2025-11-24 11:33:06.939939504 +0000 UTC m=+9647.474379966" lastFinishedPulling="2025-11-24 11:33:10.69776432 +0000 UTC m=+9651.232204782" observedRunningTime="2025-11-24 11:33:11.005809408 +0000 UTC m=+9651.540249880" watchObservedRunningTime="2025-11-24 11:33:11.015693971 +0000 UTC m=+9651.550134433" Nov 24 11:33:15 crc kubenswrapper[4944]: I1124 11:33:15.227063 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:15 crc kubenswrapper[4944]: I1124 11:33:15.227702 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:15 crc kubenswrapper[4944]: I1124 11:33:15.288964 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:16 crc kubenswrapper[4944]: I1124 11:33:16.074960 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:16 crc kubenswrapper[4944]: I1124 11:33:16.138174 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zng5q"] Nov 24 11:33:18 crc kubenswrapper[4944]: I1124 11:33:18.049192 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zng5q" podUID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" containerName="registry-server" containerID="cri-o://a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182" gracePeriod=2 Nov 24 11:33:18 crc kubenswrapper[4944]: I1124 11:33:18.556820 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:18 crc kubenswrapper[4944]: I1124 11:33:18.694306 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmtw7\" (UniqueName: \"kubernetes.io/projected/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-kube-api-access-vmtw7\") pod \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " Nov 24 11:33:18 crc kubenswrapper[4944]: I1124 11:33:18.694461 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-utilities\") pod \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " Nov 24 11:33:18 crc kubenswrapper[4944]: I1124 11:33:18.694598 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-catalog-content\") pod \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\" (UID: \"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e\") " Nov 24 11:33:18 crc kubenswrapper[4944]: I1124 11:33:18.695357 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-utilities" (OuterVolumeSpecName: "utilities") pod "8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" (UID: "8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:33:18 crc kubenswrapper[4944]: I1124 11:33:18.702158 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-kube-api-access-vmtw7" (OuterVolumeSpecName: "kube-api-access-vmtw7") pod "8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" (UID: "8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e"). InnerVolumeSpecName "kube-api-access-vmtw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:33:18 crc kubenswrapper[4944]: I1124 11:33:18.745038 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" (UID: "8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:33:18 crc kubenswrapper[4944]: I1124 11:33:18.797869 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmtw7\" (UniqueName: \"kubernetes.io/projected/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-kube-api-access-vmtw7\") on node \"crc\" DevicePath \"\"" Nov 24 11:33:18 crc kubenswrapper[4944]: I1124 11:33:18.797910 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:33:18 crc kubenswrapper[4944]: I1124 11:33:18.797928 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.060895 4944 generic.go:334] "Generic (PLEG): container finished" podID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" containerID="a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182" exitCode=0 Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.060984 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zng5q" event={"ID":"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e","Type":"ContainerDied","Data":"a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182"} Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.061011 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zng5q" Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.061078 4944 scope.go:117] "RemoveContainer" containerID="a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182" Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.061032 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zng5q" event={"ID":"8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e","Type":"ContainerDied","Data":"bacc4ee91526c501c7c848411df73b578693e9fe0b38df35a5ef1cb7758ff7c2"} Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.086728 4944 scope.go:117] "RemoveContainer" containerID="5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d" Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.102109 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zng5q"] Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.123024 4944 scope.go:117] "RemoveContainer" containerID="301cfb1f73d5fc88721b5bc804fae8ac268cb43daab6213aab9530d7c8e75c96" Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.138509 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zng5q"] Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.167319 4944 scope.go:117] "RemoveContainer" containerID="a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182" Nov 24 11:33:19 crc kubenswrapper[4944]: E1124 11:33:19.171457 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182\": container with ID starting with a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182 not found: ID does not exist" containerID="a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182" Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.171487 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182"} err="failed to get container status \"a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182\": rpc error: code = NotFound desc = could not find container \"a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182\": container with ID starting with a3ab6c9746b4d17803a7f2d10ed5069952e440ad0207fe74e5dba370a3669182 not found: ID does not exist" Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.171508 4944 scope.go:117] "RemoveContainer" containerID="5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d" Nov 24 11:33:19 crc kubenswrapper[4944]: E1124 11:33:19.173493 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d\": container with ID starting with 5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d not found: ID does not exist" containerID="5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d" Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.173547 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d"} err="failed to get container status \"5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d\": rpc error: code = NotFound desc = could not find container \"5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d\": container with ID starting with 5d5c0c6f79a97459d1ee6063c9b8d37c04988f1c5367108d1218e003ab051f7d not found: ID does not exist" Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.173577 4944 scope.go:117] "RemoveContainer" containerID="301cfb1f73d5fc88721b5bc804fae8ac268cb43daab6213aab9530d7c8e75c96" Nov 24 11:33:19 crc kubenswrapper[4944]: E1124 11:33:19.173866 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"301cfb1f73d5fc88721b5bc804fae8ac268cb43daab6213aab9530d7c8e75c96\": container with ID starting with 301cfb1f73d5fc88721b5bc804fae8ac268cb43daab6213aab9530d7c8e75c96 not found: ID does not exist" containerID="301cfb1f73d5fc88721b5bc804fae8ac268cb43daab6213aab9530d7c8e75c96" Nov 24 11:33:19 crc kubenswrapper[4944]: I1124 11:33:19.173885 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"301cfb1f73d5fc88721b5bc804fae8ac268cb43daab6213aab9530d7c8e75c96"} err="failed to get container status \"301cfb1f73d5fc88721b5bc804fae8ac268cb43daab6213aab9530d7c8e75c96\": rpc error: code = NotFound desc = could not find container \"301cfb1f73d5fc88721b5bc804fae8ac268cb43daab6213aab9530d7c8e75c96\": container with ID starting with 301cfb1f73d5fc88721b5bc804fae8ac268cb43daab6213aab9530d7c8e75c96 not found: ID does not exist" Nov 24 11:33:20 crc kubenswrapper[4944]: I1124 11:33:20.289274 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" path="/var/lib/kubelet/pods/8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e/volumes" Nov 24 11:33:23 crc kubenswrapper[4944]: I1124 11:33:23.549564 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:33:23 crc kubenswrapper[4944]: I1124 11:33:23.550198 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:33:23 crc kubenswrapper[4944]: I1124 11:33:23.550242 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 11:33:23 crc kubenswrapper[4944]: I1124 11:33:23.550904 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e2aa81f7eceb76b67b33340bde645da1f586279fcd8142d56e3e89877f97e54b"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 11:33:23 crc kubenswrapper[4944]: I1124 11:33:23.550955 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://e2aa81f7eceb76b67b33340bde645da1f586279fcd8142d56e3e89877f97e54b" gracePeriod=600 Nov 24 11:33:24 crc kubenswrapper[4944]: I1124 11:33:24.130382 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="e2aa81f7eceb76b67b33340bde645da1f586279fcd8142d56e3e89877f97e54b" exitCode=0 Nov 24 11:33:24 crc kubenswrapper[4944]: I1124 11:33:24.130426 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"e2aa81f7eceb76b67b33340bde645da1f586279fcd8142d56e3e89877f97e54b"} Nov 24 11:33:24 crc kubenswrapper[4944]: I1124 11:33:24.130760 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c"} Nov 24 11:33:24 crc kubenswrapper[4944]: I1124 11:33:24.130803 4944 scope.go:117] "RemoveContainer" containerID="90fb3409aaa997d91e288bd9257a4185e815a7fe5578cffdbe25d80badc27f20" Nov 24 11:33:31 crc kubenswrapper[4944]: I1124 11:33:31.795984 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="56ccc568-803f-4862-b9d6-f5e29d45f5b0" containerName="galera" probeResult="failure" output="command timed out" Nov 24 11:33:31 crc kubenswrapper[4944]: I1124 11:33:31.796201 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="56ccc568-803f-4862-b9d6-f5e29d45f5b0" containerName="galera" probeResult="failure" output="command timed out" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.260206 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9dnq7"] Nov 24 11:34:28 crc kubenswrapper[4944]: E1124 11:34:28.261438 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" containerName="extract-content" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.261457 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" containerName="extract-content" Nov 24 11:34:28 crc kubenswrapper[4944]: E1124 11:34:28.261487 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" containerName="registry-server" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.261496 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" containerName="registry-server" Nov 24 11:34:28 crc kubenswrapper[4944]: E1124 11:34:28.261553 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" containerName="extract-utilities" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.261562 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" containerName="extract-utilities" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.261812 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e3fc2b9-aa6c-47cc-abb5-e437be1d3e3e" containerName="registry-server" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.264017 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.274208 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9dnq7"] Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.362179 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-catalog-content\") pod \"redhat-operators-9dnq7\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.362235 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-utilities\") pod \"redhat-operators-9dnq7\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.362395 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c78ss\" (UniqueName: \"kubernetes.io/projected/21cf838a-56a1-46c3-94d0-e4876eac3cf4-kube-api-access-c78ss\") pod \"redhat-operators-9dnq7\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.465092 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-catalog-content\") pod \"redhat-operators-9dnq7\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.465160 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-utilities\") pod \"redhat-operators-9dnq7\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.465232 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c78ss\" (UniqueName: \"kubernetes.io/projected/21cf838a-56a1-46c3-94d0-e4876eac3cf4-kube-api-access-c78ss\") pod \"redhat-operators-9dnq7\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.466243 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-catalog-content\") pod \"redhat-operators-9dnq7\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.466443 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-utilities\") pod \"redhat-operators-9dnq7\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.490761 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c78ss\" (UniqueName: \"kubernetes.io/projected/21cf838a-56a1-46c3-94d0-e4876eac3cf4-kube-api-access-c78ss\") pod \"redhat-operators-9dnq7\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:28 crc kubenswrapper[4944]: I1124 11:34:28.602850 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:29 crc kubenswrapper[4944]: I1124 11:34:29.115722 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9dnq7"] Nov 24 11:34:29 crc kubenswrapper[4944]: I1124 11:34:29.860786 4944 generic.go:334] "Generic (PLEG): container finished" podID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerID="c2dcad8f918322a7607a012ff869cdfbc63a28321de58ad27f3f73d61dae0166" exitCode=0 Nov 24 11:34:29 crc kubenswrapper[4944]: I1124 11:34:29.860834 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dnq7" event={"ID":"21cf838a-56a1-46c3-94d0-e4876eac3cf4","Type":"ContainerDied","Data":"c2dcad8f918322a7607a012ff869cdfbc63a28321de58ad27f3f73d61dae0166"} Nov 24 11:34:29 crc kubenswrapper[4944]: I1124 11:34:29.861366 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dnq7" event={"ID":"21cf838a-56a1-46c3-94d0-e4876eac3cf4","Type":"ContainerStarted","Data":"98b50650c8a6d8a274ac8ecabc8795d7e51cba19eb5f27cf0aa54a2790854757"} Nov 24 11:34:31 crc kubenswrapper[4944]: I1124 11:34:31.886107 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dnq7" event={"ID":"21cf838a-56a1-46c3-94d0-e4876eac3cf4","Type":"ContainerStarted","Data":"9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1"} Nov 24 11:34:35 crc kubenswrapper[4944]: I1124 11:34:35.987513 4944 generic.go:334] "Generic (PLEG): container finished" podID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerID="9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1" exitCode=0 Nov 24 11:34:35 crc kubenswrapper[4944]: I1124 11:34:35.987677 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dnq7" event={"ID":"21cf838a-56a1-46c3-94d0-e4876eac3cf4","Type":"ContainerDied","Data":"9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1"} Nov 24 11:34:37 crc kubenswrapper[4944]: I1124 11:34:37.004395 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dnq7" event={"ID":"21cf838a-56a1-46c3-94d0-e4876eac3cf4","Type":"ContainerStarted","Data":"1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67"} Nov 24 11:34:37 crc kubenswrapper[4944]: I1124 11:34:37.069360 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9dnq7" podStartSLOduration=2.540995721 podStartE2EDuration="9.069321643s" podCreationTimestamp="2025-11-24 11:34:28 +0000 UTC" firstStartedPulling="2025-11-24 11:34:29.863692043 +0000 UTC m=+9730.398132505" lastFinishedPulling="2025-11-24 11:34:36.392017955 +0000 UTC m=+9736.926458427" observedRunningTime="2025-11-24 11:34:37.040842466 +0000 UTC m=+9737.575282938" watchObservedRunningTime="2025-11-24 11:34:37.069321643 +0000 UTC m=+9737.603762115" Nov 24 11:34:38 crc kubenswrapper[4944]: I1124 11:34:38.603406 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:38 crc kubenswrapper[4944]: I1124 11:34:38.604027 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:39 crc kubenswrapper[4944]: I1124 11:34:39.674131 4944 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9dnq7" podUID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerName="registry-server" probeResult="failure" output=< Nov 24 11:34:39 crc kubenswrapper[4944]: timeout: failed to connect service ":50051" within 1s Nov 24 11:34:39 crc kubenswrapper[4944]: > Nov 24 11:34:48 crc kubenswrapper[4944]: I1124 11:34:48.667304 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:48 crc kubenswrapper[4944]: I1124 11:34:48.733546 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:48 crc kubenswrapper[4944]: I1124 11:34:48.901264 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9dnq7"] Nov 24 11:34:50 crc kubenswrapper[4944]: I1124 11:34:50.139967 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9dnq7" podUID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerName="registry-server" containerID="cri-o://1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67" gracePeriod=2 Nov 24 11:34:50 crc kubenswrapper[4944]: I1124 11:34:50.692467 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:50 crc kubenswrapper[4944]: I1124 11:34:50.836549 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-utilities\") pod \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " Nov 24 11:34:50 crc kubenswrapper[4944]: I1124 11:34:50.836935 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c78ss\" (UniqueName: \"kubernetes.io/projected/21cf838a-56a1-46c3-94d0-e4876eac3cf4-kube-api-access-c78ss\") pod \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " Nov 24 11:34:50 crc kubenswrapper[4944]: I1124 11:34:50.836974 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-catalog-content\") pod \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\" (UID: \"21cf838a-56a1-46c3-94d0-e4876eac3cf4\") " Nov 24 11:34:50 crc kubenswrapper[4944]: I1124 11:34:50.837445 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-utilities" (OuterVolumeSpecName: "utilities") pod "21cf838a-56a1-46c3-94d0-e4876eac3cf4" (UID: "21cf838a-56a1-46c3-94d0-e4876eac3cf4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:34:50 crc kubenswrapper[4944]: I1124 11:34:50.839261 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:34:50 crc kubenswrapper[4944]: I1124 11:34:50.843359 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21cf838a-56a1-46c3-94d0-e4876eac3cf4-kube-api-access-c78ss" (OuterVolumeSpecName: "kube-api-access-c78ss") pod "21cf838a-56a1-46c3-94d0-e4876eac3cf4" (UID: "21cf838a-56a1-46c3-94d0-e4876eac3cf4"). InnerVolumeSpecName "kube-api-access-c78ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:34:50 crc kubenswrapper[4944]: I1124 11:34:50.937240 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21cf838a-56a1-46c3-94d0-e4876eac3cf4" (UID: "21cf838a-56a1-46c3-94d0-e4876eac3cf4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:34:50 crc kubenswrapper[4944]: I1124 11:34:50.945862 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c78ss\" (UniqueName: \"kubernetes.io/projected/21cf838a-56a1-46c3-94d0-e4876eac3cf4-kube-api-access-c78ss\") on node \"crc\" DevicePath \"\"" Nov 24 11:34:50 crc kubenswrapper[4944]: I1124 11:34:50.946034 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21cf838a-56a1-46c3-94d0-e4876eac3cf4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.157712 4944 generic.go:334] "Generic (PLEG): container finished" podID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerID="1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67" exitCode=0 Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.157792 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dnq7" event={"ID":"21cf838a-56a1-46c3-94d0-e4876eac3cf4","Type":"ContainerDied","Data":"1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67"} Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.157882 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9dnq7" event={"ID":"21cf838a-56a1-46c3-94d0-e4876eac3cf4","Type":"ContainerDied","Data":"98b50650c8a6d8a274ac8ecabc8795d7e51cba19eb5f27cf0aa54a2790854757"} Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.157912 4944 scope.go:117] "RemoveContainer" containerID="1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67" Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.159142 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9dnq7" Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.188600 4944 scope.go:117] "RemoveContainer" containerID="9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1" Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.205312 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9dnq7"] Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.215855 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9dnq7"] Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.233386 4944 scope.go:117] "RemoveContainer" containerID="c2dcad8f918322a7607a012ff869cdfbc63a28321de58ad27f3f73d61dae0166" Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.278578 4944 scope.go:117] "RemoveContainer" containerID="1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67" Nov 24 11:34:51 crc kubenswrapper[4944]: E1124 11:34:51.279685 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67\": container with ID starting with 1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67 not found: ID does not exist" containerID="1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67" Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.279728 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67"} err="failed to get container status \"1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67\": rpc error: code = NotFound desc = could not find container \"1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67\": container with ID starting with 1398b4c43701574ca69c6b4738ff4a53039d828be580994f7246c5112a8efb67 not found: ID does not exist" Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.279756 4944 scope.go:117] "RemoveContainer" containerID="9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1" Nov 24 11:34:51 crc kubenswrapper[4944]: E1124 11:34:51.280979 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1\": container with ID starting with 9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1 not found: ID does not exist" containerID="9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1" Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.281004 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1"} err="failed to get container status \"9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1\": rpc error: code = NotFound desc = could not find container \"9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1\": container with ID starting with 9c8d08549d3664050e25f028b36ce4399be2d3ed6667e56e941d31f3207908d1 not found: ID does not exist" Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.281020 4944 scope.go:117] "RemoveContainer" containerID="c2dcad8f918322a7607a012ff869cdfbc63a28321de58ad27f3f73d61dae0166" Nov 24 11:34:51 crc kubenswrapper[4944]: E1124 11:34:51.281473 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2dcad8f918322a7607a012ff869cdfbc63a28321de58ad27f3f73d61dae0166\": container with ID starting with c2dcad8f918322a7607a012ff869cdfbc63a28321de58ad27f3f73d61dae0166 not found: ID does not exist" containerID="c2dcad8f918322a7607a012ff869cdfbc63a28321de58ad27f3f73d61dae0166" Nov 24 11:34:51 crc kubenswrapper[4944]: I1124 11:34:51.281506 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2dcad8f918322a7607a012ff869cdfbc63a28321de58ad27f3f73d61dae0166"} err="failed to get container status \"c2dcad8f918322a7607a012ff869cdfbc63a28321de58ad27f3f73d61dae0166\": rpc error: code = NotFound desc = could not find container \"c2dcad8f918322a7607a012ff869cdfbc63a28321de58ad27f3f73d61dae0166\": container with ID starting with c2dcad8f918322a7607a012ff869cdfbc63a28321de58ad27f3f73d61dae0166 not found: ID does not exist" Nov 24 11:34:52 crc kubenswrapper[4944]: I1124 11:34:52.288330 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" path="/var/lib/kubelet/pods/21cf838a-56a1-46c3-94d0-e4876eac3cf4/volumes" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.341204 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9rrzt"] Nov 24 11:34:54 crc kubenswrapper[4944]: E1124 11:34:54.342812 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerName="extract-content" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.342828 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerName="extract-content" Nov 24 11:34:54 crc kubenswrapper[4944]: E1124 11:34:54.342865 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerName="registry-server" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.342872 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerName="registry-server" Nov 24 11:34:54 crc kubenswrapper[4944]: E1124 11:34:54.342886 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerName="extract-utilities" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.342893 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerName="extract-utilities" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.343199 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="21cf838a-56a1-46c3-94d0-e4876eac3cf4" containerName="registry-server" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.345006 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.381324 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9rrzt"] Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.418992 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kzwm\" (UniqueName: \"kubernetes.io/projected/d12c6fc4-dfe2-4633-b681-8bee73f872e3-kube-api-access-8kzwm\") pod \"community-operators-9rrzt\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.420470 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-catalog-content\") pod \"community-operators-9rrzt\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.420618 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-utilities\") pod \"community-operators-9rrzt\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.522873 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-utilities\") pod \"community-operators-9rrzt\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.523205 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kzwm\" (UniqueName: \"kubernetes.io/projected/d12c6fc4-dfe2-4633-b681-8bee73f872e3-kube-api-access-8kzwm\") pod \"community-operators-9rrzt\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.523383 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-utilities\") pod \"community-operators-9rrzt\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.523401 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-catalog-content\") pod \"community-operators-9rrzt\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.523804 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-catalog-content\") pod \"community-operators-9rrzt\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.544163 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kzwm\" (UniqueName: \"kubernetes.io/projected/d12c6fc4-dfe2-4633-b681-8bee73f872e3-kube-api-access-8kzwm\") pod \"community-operators-9rrzt\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:34:54 crc kubenswrapper[4944]: I1124 11:34:54.706730 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:34:55 crc kubenswrapper[4944]: I1124 11:34:55.285698 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9rrzt"] Nov 24 11:34:56 crc kubenswrapper[4944]: I1124 11:34:56.215230 4944 generic.go:334] "Generic (PLEG): container finished" podID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" containerID="ac811e383c6b7ae3a56b5956bb8274ea7e3cc3c11f4930cb2340c52595a36b62" exitCode=0 Nov 24 11:34:56 crc kubenswrapper[4944]: I1124 11:34:56.215562 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rrzt" event={"ID":"d12c6fc4-dfe2-4633-b681-8bee73f872e3","Type":"ContainerDied","Data":"ac811e383c6b7ae3a56b5956bb8274ea7e3cc3c11f4930cb2340c52595a36b62"} Nov 24 11:34:56 crc kubenswrapper[4944]: I1124 11:34:56.215598 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rrzt" event={"ID":"d12c6fc4-dfe2-4633-b681-8bee73f872e3","Type":"ContainerStarted","Data":"f6d866f9ec509a1bb3f30f01590f67a5a80835ce4144dfd8aed85f9320dfac64"} Nov 24 11:34:58 crc kubenswrapper[4944]: I1124 11:34:58.236149 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rrzt" event={"ID":"d12c6fc4-dfe2-4633-b681-8bee73f872e3","Type":"ContainerStarted","Data":"35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1"} Nov 24 11:34:59 crc kubenswrapper[4944]: I1124 11:34:59.248354 4944 generic.go:334] "Generic (PLEG): container finished" podID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" containerID="35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1" exitCode=0 Nov 24 11:34:59 crc kubenswrapper[4944]: I1124 11:34:59.248414 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rrzt" event={"ID":"d12c6fc4-dfe2-4633-b681-8bee73f872e3","Type":"ContainerDied","Data":"35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1"} Nov 24 11:35:00 crc kubenswrapper[4944]: I1124 11:35:00.257918 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rrzt" event={"ID":"d12c6fc4-dfe2-4633-b681-8bee73f872e3","Type":"ContainerStarted","Data":"75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0"} Nov 24 11:35:00 crc kubenswrapper[4944]: I1124 11:35:00.286243 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9rrzt" podStartSLOduration=2.81026848 podStartE2EDuration="6.286223774s" podCreationTimestamp="2025-11-24 11:34:54 +0000 UTC" firstStartedPulling="2025-11-24 11:34:56.221169445 +0000 UTC m=+9756.755609907" lastFinishedPulling="2025-11-24 11:34:59.697124739 +0000 UTC m=+9760.231565201" observedRunningTime="2025-11-24 11:35:00.281471424 +0000 UTC m=+9760.815911886" watchObservedRunningTime="2025-11-24 11:35:00.286223774 +0000 UTC m=+9760.820664256" Nov 24 11:35:04 crc kubenswrapper[4944]: I1124 11:35:04.706896 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:35:04 crc kubenswrapper[4944]: I1124 11:35:04.708513 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:35:04 crc kubenswrapper[4944]: I1124 11:35:04.757644 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:35:05 crc kubenswrapper[4944]: I1124 11:35:05.361508 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:35:05 crc kubenswrapper[4944]: I1124 11:35:05.416710 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9rrzt"] Nov 24 11:35:07 crc kubenswrapper[4944]: I1124 11:35:07.333678 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9rrzt" podUID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" containerName="registry-server" containerID="cri-o://75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0" gracePeriod=2 Nov 24 11:35:07 crc kubenswrapper[4944]: I1124 11:35:07.873394 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:35:07 crc kubenswrapper[4944]: I1124 11:35:07.981951 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kzwm\" (UniqueName: \"kubernetes.io/projected/d12c6fc4-dfe2-4633-b681-8bee73f872e3-kube-api-access-8kzwm\") pod \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " Nov 24 11:35:07 crc kubenswrapper[4944]: I1124 11:35:07.982104 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-utilities\") pod \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " Nov 24 11:35:07 crc kubenswrapper[4944]: I1124 11:35:07.982403 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-catalog-content\") pod \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\" (UID: \"d12c6fc4-dfe2-4633-b681-8bee73f872e3\") " Nov 24 11:35:07 crc kubenswrapper[4944]: I1124 11:35:07.986163 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-utilities" (OuterVolumeSpecName: "utilities") pod "d12c6fc4-dfe2-4633-b681-8bee73f872e3" (UID: "d12c6fc4-dfe2-4633-b681-8bee73f872e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.001885 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d12c6fc4-dfe2-4633-b681-8bee73f872e3-kube-api-access-8kzwm" (OuterVolumeSpecName: "kube-api-access-8kzwm") pod "d12c6fc4-dfe2-4633-b681-8bee73f872e3" (UID: "d12c6fc4-dfe2-4633-b681-8bee73f872e3"). InnerVolumeSpecName "kube-api-access-8kzwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.051491 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d12c6fc4-dfe2-4633-b681-8bee73f872e3" (UID: "d12c6fc4-dfe2-4633-b681-8bee73f872e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.085633 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kzwm\" (UniqueName: \"kubernetes.io/projected/d12c6fc4-dfe2-4633-b681-8bee73f872e3-kube-api-access-8kzwm\") on node \"crc\" DevicePath \"\"" Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.085676 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.085687 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d12c6fc4-dfe2-4633-b681-8bee73f872e3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.346012 4944 generic.go:334] "Generic (PLEG): container finished" podID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" containerID="75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0" exitCode=0 Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.346062 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rrzt" event={"ID":"d12c6fc4-dfe2-4633-b681-8bee73f872e3","Type":"ContainerDied","Data":"75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0"} Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.346091 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rrzt" event={"ID":"d12c6fc4-dfe2-4633-b681-8bee73f872e3","Type":"ContainerDied","Data":"f6d866f9ec509a1bb3f30f01590f67a5a80835ce4144dfd8aed85f9320dfac64"} Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.346108 4944 scope.go:117] "RemoveContainer" containerID="75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0" Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.346262 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rrzt" Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.382183 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9rrzt"] Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.383408 4944 scope.go:117] "RemoveContainer" containerID="35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1" Nov 24 11:35:08 crc kubenswrapper[4944]: I1124 11:35:08.393740 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9rrzt"] Nov 24 11:35:09 crc kubenswrapper[4944]: I1124 11:35:09.012076 4944 scope.go:117] "RemoveContainer" containerID="ac811e383c6b7ae3a56b5956bb8274ea7e3cc3c11f4930cb2340c52595a36b62" Nov 24 11:35:09 crc kubenswrapper[4944]: I1124 11:35:09.122232 4944 scope.go:117] "RemoveContainer" containerID="75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0" Nov 24 11:35:09 crc kubenswrapper[4944]: E1124 11:35:09.125905 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0\": container with ID starting with 75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0 not found: ID does not exist" containerID="75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0" Nov 24 11:35:09 crc kubenswrapper[4944]: I1124 11:35:09.125944 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0"} err="failed to get container status \"75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0\": rpc error: code = NotFound desc = could not find container \"75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0\": container with ID starting with 75d4df0e902dd6ac56c07e824f6d5286594e847f87ebd927354f86160d971fe0 not found: ID does not exist" Nov 24 11:35:09 crc kubenswrapper[4944]: I1124 11:35:09.125966 4944 scope.go:117] "RemoveContainer" containerID="35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1" Nov 24 11:35:09 crc kubenswrapper[4944]: E1124 11:35:09.127680 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1\": container with ID starting with 35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1 not found: ID does not exist" containerID="35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1" Nov 24 11:35:09 crc kubenswrapper[4944]: I1124 11:35:09.127778 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1"} err="failed to get container status \"35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1\": rpc error: code = NotFound desc = could not find container \"35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1\": container with ID starting with 35698ca8fe22c0ea7b48ed214c6be9d8af7e3a105cdf0948d98e940cca2eecb1 not found: ID does not exist" Nov 24 11:35:09 crc kubenswrapper[4944]: I1124 11:35:09.127853 4944 scope.go:117] "RemoveContainer" containerID="ac811e383c6b7ae3a56b5956bb8274ea7e3cc3c11f4930cb2340c52595a36b62" Nov 24 11:35:09 crc kubenswrapper[4944]: E1124 11:35:09.128416 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac811e383c6b7ae3a56b5956bb8274ea7e3cc3c11f4930cb2340c52595a36b62\": container with ID starting with ac811e383c6b7ae3a56b5956bb8274ea7e3cc3c11f4930cb2340c52595a36b62 not found: ID does not exist" containerID="ac811e383c6b7ae3a56b5956bb8274ea7e3cc3c11f4930cb2340c52595a36b62" Nov 24 11:35:09 crc kubenswrapper[4944]: I1124 11:35:09.128517 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac811e383c6b7ae3a56b5956bb8274ea7e3cc3c11f4930cb2340c52595a36b62"} err="failed to get container status \"ac811e383c6b7ae3a56b5956bb8274ea7e3cc3c11f4930cb2340c52595a36b62\": rpc error: code = NotFound desc = could not find container \"ac811e383c6b7ae3a56b5956bb8274ea7e3cc3c11f4930cb2340c52595a36b62\": container with ID starting with ac811e383c6b7ae3a56b5956bb8274ea7e3cc3c11f4930cb2340c52595a36b62 not found: ID does not exist" Nov 24 11:35:10 crc kubenswrapper[4944]: I1124 11:35:10.049272 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_be3827d1-c5b1-4048-af2f-e2a2be15a396/init-config-reloader/0.log" Nov 24 11:35:10 crc kubenswrapper[4944]: I1124 11:35:10.272759 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_be3827d1-c5b1-4048-af2f-e2a2be15a396/init-config-reloader/0.log" Nov 24 11:35:10 crc kubenswrapper[4944]: I1124 11:35:10.276276 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_be3827d1-c5b1-4048-af2f-e2a2be15a396/alertmanager/0.log" Nov 24 11:35:10 crc kubenswrapper[4944]: I1124 11:35:10.291554 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" path="/var/lib/kubelet/pods/d12c6fc4-dfe2-4633-b681-8bee73f872e3/volumes" Nov 24 11:35:10 crc kubenswrapper[4944]: I1124 11:35:10.348766 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_be3827d1-c5b1-4048-af2f-e2a2be15a396/config-reloader/0.log" Nov 24 11:35:10 crc kubenswrapper[4944]: I1124 11:35:10.542012 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8e9d68ad-ee1d-49f3-a387-62129d8b3a26/aodh-api/0.log" Nov 24 11:35:10 crc kubenswrapper[4944]: I1124 11:35:10.620328 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8e9d68ad-ee1d-49f3-a387-62129d8b3a26/aodh-evaluator/0.log" Nov 24 11:35:10 crc kubenswrapper[4944]: I1124 11:35:10.690362 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8e9d68ad-ee1d-49f3-a387-62129d8b3a26/aodh-listener/0.log" Nov 24 11:35:10 crc kubenswrapper[4944]: I1124 11:35:10.785857 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8e9d68ad-ee1d-49f3-a387-62129d8b3a26/aodh-notifier/0.log" Nov 24 11:35:10 crc kubenswrapper[4944]: I1124 11:35:10.854536 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-668df884fb-fs274_9836c3a7-f5d2-468c-a3a7-03882013755a/barbican-api/0.log" Nov 24 11:35:11 crc kubenswrapper[4944]: I1124 11:35:11.472720 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-cfc58c9fb-l9b6d_0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4/barbican-keystone-listener/0.log" Nov 24 11:35:11 crc kubenswrapper[4944]: I1124 11:35:11.476522 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-cfc58c9fb-l9b6d_0c84b0f6-464d-4ab3-8fe5-19bb31a2a1b4/barbican-keystone-listener-log/0.log" Nov 24 11:35:11 crc kubenswrapper[4944]: I1124 11:35:11.545802 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-668df884fb-fs274_9836c3a7-f5d2-468c-a3a7-03882013755a/barbican-api-log/0.log" Nov 24 11:35:11 crc kubenswrapper[4944]: I1124 11:35:11.659328 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-69767489f-22wtk_77bc0951-6f9e-4a4f-be92-fadde903228c/barbican-worker/0.log" Nov 24 11:35:11 crc kubenswrapper[4944]: I1124 11:35:11.703686 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-69767489f-22wtk_77bc0951-6f9e-4a4f-be92-fadde903228c/barbican-worker-log/0.log" Nov 24 11:35:12 crc kubenswrapper[4944]: I1124 11:35:12.158689 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-9pwb2_37b4d7f7-90d4-466c-92a1-dcd83eba1306/bootstrap-openstack-openstack-cell1/0.log" Nov 24 11:35:12 crc kubenswrapper[4944]: I1124 11:35:12.193949 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ec6f15bc-0236-428d-9609-b1b24f963628/ceilometer-central-agent/0.log" Nov 24 11:35:12 crc kubenswrapper[4944]: I1124 11:35:12.250964 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ec6f15bc-0236-428d-9609-b1b24f963628/ceilometer-notification-agent/0.log" Nov 24 11:35:12 crc kubenswrapper[4944]: I1124 11:35:12.346604 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ec6f15bc-0236-428d-9609-b1b24f963628/proxy-httpd/0.log" Nov 24 11:35:12 crc kubenswrapper[4944]: I1124 11:35:12.433418 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ec6f15bc-0236-428d-9609-b1b24f963628/sg-core/0.log" Nov 24 11:35:12 crc kubenswrapper[4944]: I1124 11:35:12.519091 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-mw6c6_23ab44ae-eb0d-47a7-b321-7b53ab21ddff/ceph-client-openstack-openstack-cell1/0.log" Nov 24 11:35:12 crc kubenswrapper[4944]: I1124 11:35:12.720284 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_94910134-a6a9-4b2c-b29d-24098a44d4a3/cinder-api-log/0.log" Nov 24 11:35:12 crc kubenswrapper[4944]: I1124 11:35:12.765709 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_94910134-a6a9-4b2c-b29d-24098a44d4a3/cinder-api/0.log" Nov 24 11:35:12 crc kubenswrapper[4944]: I1124 11:35:12.920432 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_d4885864-e680-460d-9145-edb86bfabf41/cinder-backup/0.log" Nov 24 11:35:13 crc kubenswrapper[4944]: I1124 11:35:13.005561 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_d4885864-e680-460d-9145-edb86bfabf41/probe/0.log" Nov 24 11:35:13 crc kubenswrapper[4944]: I1124 11:35:13.153149 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e92855b-657c-4b26-a3b0-6916f4f2cb81/cinder-scheduler/0.log" Nov 24 11:35:13 crc kubenswrapper[4944]: I1124 11:35:13.198232 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e92855b-657c-4b26-a3b0-6916f4f2cb81/probe/0.log" Nov 24 11:35:13 crc kubenswrapper[4944]: I1124 11:35:13.380119 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_07dbc464-82a3-4762-a75b-1b4b7a236be7/cinder-volume/0.log" Nov 24 11:35:13 crc kubenswrapper[4944]: I1124 11:35:13.452810 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_07dbc464-82a3-4762-a75b-1b4b7a236be7/probe/0.log" Nov 24 11:35:13 crc kubenswrapper[4944]: I1124 11:35:13.473559 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-sxf2s_f09923cd-e53d-4abd-a2e6-1bec30789bbd/configure-network-openstack-openstack-cell1/0.log" Nov 24 11:35:13 crc kubenswrapper[4944]: I1124 11:35:13.620086 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-vp6h7_9b3afe10-ef54-431f-91b1-b05e0900d544/configure-os-openstack-openstack-cell1/0.log" Nov 24 11:35:13 crc kubenswrapper[4944]: I1124 11:35:13.724297 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5b454bb789-4gtgk_8af65607-21e8-4d80-8d5f-12faa2e45caf/init/0.log" Nov 24 11:35:13 crc kubenswrapper[4944]: I1124 11:35:13.905383 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5b454bb789-4gtgk_8af65607-21e8-4d80-8d5f-12faa2e45caf/init/0.log" Nov 24 11:35:13 crc kubenswrapper[4944]: I1124 11:35:13.979000 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-7tdq4_1b22e4e6-4037-4c1d-a2b1-b428efd1f470/download-cache-openstack-openstack-cell1/0.log" Nov 24 11:35:13 crc kubenswrapper[4944]: I1124 11:35:13.981773 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5b454bb789-4gtgk_8af65607-21e8-4d80-8d5f-12faa2e45caf/dnsmasq-dns/0.log" Nov 24 11:35:14 crc kubenswrapper[4944]: I1124 11:35:14.189330 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_942ae185-df7d-4585-995a-dc253806d0fe/glance-log/0.log" Nov 24 11:35:14 crc kubenswrapper[4944]: I1124 11:35:14.213183 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_942ae185-df7d-4585-995a-dc253806d0fe/glance-httpd/0.log" Nov 24 11:35:14 crc kubenswrapper[4944]: I1124 11:35:14.321698 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d2765b02-6d6d-42c2-bbf1-92ae32ef6662/glance-httpd/0.log" Nov 24 11:35:14 crc kubenswrapper[4944]: I1124 11:35:14.353029 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d2765b02-6d6d-42c2-bbf1-92ae32ef6662/glance-log/0.log" Nov 24 11:35:14 crc kubenswrapper[4944]: I1124 11:35:14.667234 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-6d6c5dbc76-jsn6f_d05697e4-6cc0-4cc0-a5a9-9e2c3d8d6ec9/heat-cfnapi/0.log" Nov 24 11:35:14 crc kubenswrapper[4944]: I1124 11:35:14.683899 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-5987fb6488-ftkqz_7f750eb4-b853-445a-acd3-dece22f6bcf3/heat-api/0.log" Nov 24 11:35:14 crc kubenswrapper[4944]: I1124 11:35:14.823098 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-9544b586c-rkdrm_430c37b1-3658-4472-aa78-ac54f934d9fb/heat-engine/0.log" Nov 24 11:35:15 crc kubenswrapper[4944]: I1124 11:35:15.063800 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-74876d679c-sbxz5_946e72b9-d3e9-4823-8b87-2049e4a107ac/horizon/0.log" Nov 24 11:35:15 crc kubenswrapper[4944]: I1124 11:35:15.100364 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-gw8rw_bde6f616-2438-4d9d-8c87-a0f8257ab2fb/install-certs-openstack-openstack-cell1/0.log" Nov 24 11:35:15 crc kubenswrapper[4944]: I1124 11:35:15.115000 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-74876d679c-sbxz5_946e72b9-d3e9-4823-8b87-2049e4a107ac/horizon-log/0.log" Nov 24 11:35:15 crc kubenswrapper[4944]: I1124 11:35:15.317937 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-89pgd_3317708a-6513-47e2-b70f-b4137e7e25d1/install-os-openstack-openstack-cell1/0.log" Nov 24 11:35:15 crc kubenswrapper[4944]: I1124 11:35:15.574356 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-8944d6bb7-str5n_aa76b57b-2f12-4a98-a477-25b528e65e55/keystone-api/0.log" Nov 24 11:35:15 crc kubenswrapper[4944]: I1124 11:35:15.575748 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29399701-jl9qm_72f27aa3-7622-4e53-bef1-89f30315ceb3/keystone-cron/0.log" Nov 24 11:35:15 crc kubenswrapper[4944]: I1124 11:35:15.692633 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_afd374c3-c2ec-4c15-a821-9223bd6e7e2d/kube-state-metrics/0.log" Nov 24 11:35:15 crc kubenswrapper[4944]: I1124 11:35:15.827760 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-2nlwg_2436e674-f08d-40d2-9c86-79e807205863/libvirt-openstack-openstack-cell1/0.log" Nov 24 11:35:16 crc kubenswrapper[4944]: I1124 11:35:16.007983 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8c8c89d3-ef27-4d5c-b083-20d868aa86a9/manila-api-log/0.log" Nov 24 11:35:16 crc kubenswrapper[4944]: I1124 11:35:16.009394 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_8c8c89d3-ef27-4d5c-b083-20d868aa86a9/manila-api/0.log" Nov 24 11:35:16 crc kubenswrapper[4944]: I1124 11:35:16.134861 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_eb4408f3-0300-4601-b866-94c5ddb146eb/probe/0.log" Nov 24 11:35:16 crc kubenswrapper[4944]: I1124 11:35:16.199118 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_eb4408f3-0300-4601-b866-94c5ddb146eb/manila-scheduler/0.log" Nov 24 11:35:16 crc kubenswrapper[4944]: I1124 11:35:16.281098 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d23e50d9-f904-4ee1-8ab4-2731ee07e643/manila-share/0.log" Nov 24 11:35:16 crc kubenswrapper[4944]: I1124 11:35:16.378231 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d23e50d9-f904-4ee1-8ab4-2731ee07e643/probe/0.log" Nov 24 11:35:16 crc kubenswrapper[4944]: I1124 11:35:16.699840 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-557c4b8bbf-f4p6l_55f037d2-e065-4810-92a4-92ed6753b333/neutron-httpd/0.log" Nov 24 11:35:16 crc kubenswrapper[4944]: I1124 11:35:16.704570 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-557c4b8bbf-f4p6l_55f037d2-e065-4810-92a4-92ed6753b333/neutron-api/0.log" Nov 24 11:35:17 crc kubenswrapper[4944]: I1124 11:35:17.047924 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-mhrzw_3dea0418-b412-4bb6-856c-f34de24377a1/neutron-dhcp-openstack-openstack-cell1/0.log" Nov 24 11:35:17 crc kubenswrapper[4944]: I1124 11:35:17.157474 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-sp7hl_ed64daec-13f6-4d5f-8f9b-f1cd983d5727/neutron-metadata-openstack-openstack-cell1/0.log" Nov 24 11:35:17 crc kubenswrapper[4944]: I1124 11:35:17.312792 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-k5pgj_20c0c984-e49b-42f6-b3ee-385aa90befc5/neutron-sriov-openstack-openstack-cell1/0.log" Nov 24 11:35:17 crc kubenswrapper[4944]: I1124 11:35:17.525098 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a521fc51-f8b9-426d-a066-735ce6868d6f/nova-api-api/0.log" Nov 24 11:35:17 crc kubenswrapper[4944]: I1124 11:35:17.635347 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_a521fc51-f8b9-426d-a066-735ce6868d6f/nova-api-log/0.log" Nov 24 11:35:17 crc kubenswrapper[4944]: I1124 11:35:17.781756 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_2142325c-7d11-49dd-ba85-e38b4da99c40/nova-cell0-conductor-conductor/0.log" Nov 24 11:35:17 crc kubenswrapper[4944]: I1124 11:35:17.890517 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_bb7c0120-2a92-4a79-b544-5022417126c3/nova-cell1-conductor-conductor/0.log" Nov 24 11:35:18 crc kubenswrapper[4944]: I1124 11:35:18.111180 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_3c726090-4f6f-40db-a2c2-6217339f122d/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 11:35:18 crc kubenswrapper[4944]: I1124 11:35:18.221008 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellq4kwp_bf413e07-7450-4eb5-8ff3-3950c9e0d44a/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Nov 24 11:35:18 crc kubenswrapper[4944]: I1124 11:35:18.471856 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-r796f_453bfabf-cd8c-46b1-b036-a1e16f1cf2ef/nova-cell1-openstack-openstack-cell1/0.log" Nov 24 11:35:18 crc kubenswrapper[4944]: I1124 11:35:18.549394 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a8da4376-b2e8-477b-b088-110d4c143a06/nova-metadata-metadata/0.log" Nov 24 11:35:18 crc kubenswrapper[4944]: I1124 11:35:18.644625 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a8da4376-b2e8-477b-b088-110d4c143a06/nova-metadata-log/0.log" Nov 24 11:35:18 crc kubenswrapper[4944]: I1124 11:35:18.795166 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d6564735-987c-4999-a5c2-d13e06e59f8a/nova-scheduler-scheduler/0.log" Nov 24 11:35:18 crc kubenswrapper[4944]: I1124 11:35:18.866501 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-75588859b8-79hh7_3012a4dd-0211-4ada-a25c-234d280caa1a/init/0.log" Nov 24 11:35:19 crc kubenswrapper[4944]: I1124 11:35:19.165896 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-75588859b8-79hh7_3012a4dd-0211-4ada-a25c-234d280caa1a/init/0.log" Nov 24 11:35:19 crc kubenswrapper[4944]: I1124 11:35:19.195177 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-75588859b8-79hh7_3012a4dd-0211-4ada-a25c-234d280caa1a/octavia-api-provider-agent/0.log" Nov 24 11:35:19 crc kubenswrapper[4944]: I1124 11:35:19.390933 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-85x5v_80b0bc60-ac8c-440b-9323-1090e5121d0a/init/0.log" Nov 24 11:35:19 crc kubenswrapper[4944]: I1124 11:35:19.417256 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-75588859b8-79hh7_3012a4dd-0211-4ada-a25c-234d280caa1a/octavia-api/0.log" Nov 24 11:35:19 crc kubenswrapper[4944]: I1124 11:35:19.570955 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-85x5v_80b0bc60-ac8c-440b-9323-1090e5121d0a/init/0.log" Nov 24 11:35:19 crc kubenswrapper[4944]: I1124 11:35:19.663082 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-85x5v_80b0bc60-ac8c-440b-9323-1090e5121d0a/octavia-healthmanager/0.log" Nov 24 11:35:19 crc kubenswrapper[4944]: I1124 11:35:19.716331 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-8mfrk_1352ff2a-e445-4c7c-930d-6356a19543eb/init/0.log" Nov 24 11:35:20 crc kubenswrapper[4944]: I1124 11:35:20.421743 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-8mfrk_1352ff2a-e445-4c7c-930d-6356a19543eb/init/0.log" Nov 24 11:35:20 crc kubenswrapper[4944]: I1124 11:35:20.447149 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-m8c9f_36c49dd9-e55b-44f7-9dd6-50ec44c9fde2/init/0.log" Nov 24 11:35:20 crc kubenswrapper[4944]: I1124 11:35:20.534668 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-8mfrk_1352ff2a-e445-4c7c-930d-6356a19543eb/octavia-housekeeping/0.log" Nov 24 11:35:20 crc kubenswrapper[4944]: I1124 11:35:20.721579 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-m8c9f_36c49dd9-e55b-44f7-9dd6-50ec44c9fde2/octavia-rsyslog/0.log" Nov 24 11:35:20 crc kubenswrapper[4944]: I1124 11:35:20.733916 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-m8c9f_36c49dd9-e55b-44f7-9dd6-50ec44c9fde2/init/0.log" Nov 24 11:35:20 crc kubenswrapper[4944]: I1124 11:35:20.820164 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-q5tb6_a22f171c-b444-4735-8c78-8ee9f41702a3/init/0.log" Nov 24 11:35:21 crc kubenswrapper[4944]: I1124 11:35:21.107619 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-q5tb6_a22f171c-b444-4735-8c78-8ee9f41702a3/init/0.log" Nov 24 11:35:21 crc kubenswrapper[4944]: I1124 11:35:21.125649 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_56ccc568-803f-4862-b9d6-f5e29d45f5b0/mysql-bootstrap/0.log" Nov 24 11:35:21 crc kubenswrapper[4944]: I1124 11:35:21.176967 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-q5tb6_a22f171c-b444-4735-8c78-8ee9f41702a3/octavia-worker/0.log" Nov 24 11:35:21 crc kubenswrapper[4944]: I1124 11:35:21.315496 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_56ccc568-803f-4862-b9d6-f5e29d45f5b0/mysql-bootstrap/0.log" Nov 24 11:35:21 crc kubenswrapper[4944]: I1124 11:35:21.441141 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d94b6e3-a5b3-40dd-8346-97a1f5ece983/mysql-bootstrap/0.log" Nov 24 11:35:21 crc kubenswrapper[4944]: I1124 11:35:21.455195 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_56ccc568-803f-4862-b9d6-f5e29d45f5b0/galera/0.log" Nov 24 11:35:21 crc kubenswrapper[4944]: I1124 11:35:21.672206 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d94b6e3-a5b3-40dd-8346-97a1f5ece983/mysql-bootstrap/0.log" Nov 24 11:35:21 crc kubenswrapper[4944]: I1124 11:35:21.739795 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1d94b6e3-a5b3-40dd-8346-97a1f5ece983/galera/0.log" Nov 24 11:35:22 crc kubenswrapper[4944]: I1124 11:35:22.283032 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_3602dfbe-14d8-4dca-a2f7-3f645acc52ad/openstackclient/0.log" Nov 24 11:35:22 crc kubenswrapper[4944]: I1124 11:35:22.349598 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-dlb49_eec5069e-41ab-4c69-b297-7b2309dc8099/ovn-controller/0.log" Nov 24 11:35:22 crc kubenswrapper[4944]: I1124 11:35:22.488629 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-cx2r8_33be9ed2-a523-499b-86c5-22d3953fe9dd/openstack-network-exporter/0.log" Nov 24 11:35:22 crc kubenswrapper[4944]: I1124 11:35:22.578576 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ps6tm_4fbacbdb-4237-4418-8bf4-54fefad49124/ovsdb-server-init/0.log" Nov 24 11:35:22 crc kubenswrapper[4944]: I1124 11:35:22.792088 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ps6tm_4fbacbdb-4237-4418-8bf4-54fefad49124/ovsdb-server/0.log" Nov 24 11:35:22 crc kubenswrapper[4944]: I1124 11:35:22.807086 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ps6tm_4fbacbdb-4237-4418-8bf4-54fefad49124/ovsdb-server-init/0.log" Nov 24 11:35:22 crc kubenswrapper[4944]: I1124 11:35:22.834489 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ps6tm_4fbacbdb-4237-4418-8bf4-54fefad49124/ovs-vswitchd/0.log" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.030399 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_561e45c8-64b9-4eef-a8a9-b1620e1d03b1/ovn-northd/0.log" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.045303 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_561e45c8-64b9-4eef-a8a9-b1620e1d03b1/openstack-network-exporter/0.log" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.260983 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_198d2038-005c-4e4d-9f85-dfa252b98986/openstack-network-exporter/0.log" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.275866 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-xmptp_1362e83f-d5d5-43c2-bba1-21fa8ec0833e/ovn-openstack-openstack-cell1/0.log" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.352894 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_198d2038-005c-4e4d-9f85-dfa252b98986/ovsdbserver-nb/0.log" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.520140 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30/openstack-network-exporter/0.log" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.548238 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.548292 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.577213 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_9fd1a0eb-c171-4cfd-8048-a8b22fc1ce30/ovsdbserver-nb/0.log" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.762890 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_65f35548-9555-4344-8f6a-0fe24432e01d/openstack-network-exporter/0.log" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.849274 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1543a026-7eaa-4a99-9cee-7cf077de95d0/openstack-network-exporter/0.log" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.866740 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_65f35548-9555-4344-8f6a-0fe24432e01d/ovsdbserver-nb/0.log" Nov 24 11:35:23 crc kubenswrapper[4944]: I1124 11:35:23.951764 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1543a026-7eaa-4a99-9cee-7cf077de95d0/ovsdbserver-sb/0.log" Nov 24 11:35:24 crc kubenswrapper[4944]: I1124 11:35:24.062095 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_6db10533-47f7-48be-b9e8-2be4aa70b7c3/openstack-network-exporter/0.log" Nov 24 11:35:24 crc kubenswrapper[4944]: I1124 11:35:24.078893 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_6db10533-47f7-48be-b9e8-2be4aa70b7c3/ovsdbserver-sb/0.log" Nov 24 11:35:24 crc kubenswrapper[4944]: I1124 11:35:24.306240 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_92de3bdd-72e5-4b71-9d2a-6279b8542747/openstack-network-exporter/0.log" Nov 24 11:35:24 crc kubenswrapper[4944]: I1124 11:35:24.383345 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_92de3bdd-72e5-4b71-9d2a-6279b8542747/ovsdbserver-sb/0.log" Nov 24 11:35:24 crc kubenswrapper[4944]: I1124 11:35:24.719705 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-774d97dd4-ks9d9_05fc0069-b41d-4157-8910-484784282a88/placement-api/0.log" Nov 24 11:35:24 crc kubenswrapper[4944]: I1124 11:35:24.830344 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-774d97dd4-ks9d9_05fc0069-b41d-4157-8910-484784282a88/placement-log/0.log" Nov 24 11:35:24 crc kubenswrapper[4944]: I1124 11:35:24.879194 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cpj46f_35e9b789-d140-4c8e-af16-8a706a439e19/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Nov 24 11:35:25 crc kubenswrapper[4944]: I1124 11:35:25.087399 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_230aefa5-9e7d-4c8f-8d3e-dd7349413f28/init-config-reloader/0.log" Nov 24 11:35:25 crc kubenswrapper[4944]: I1124 11:35:25.328344 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_230aefa5-9e7d-4c8f-8d3e-dd7349413f28/thanos-sidecar/0.log" Nov 24 11:35:25 crc kubenswrapper[4944]: I1124 11:35:25.336144 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_230aefa5-9e7d-4c8f-8d3e-dd7349413f28/prometheus/0.log" Nov 24 11:35:25 crc kubenswrapper[4944]: I1124 11:35:25.340042 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_230aefa5-9e7d-4c8f-8d3e-dd7349413f28/config-reloader/0.log" Nov 24 11:35:25 crc kubenswrapper[4944]: I1124 11:35:25.350882 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_230aefa5-9e7d-4c8f-8d3e-dd7349413f28/init-config-reloader/0.log" Nov 24 11:35:25 crc kubenswrapper[4944]: I1124 11:35:25.562257 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_15ce2947-cdd4-4b26-99c1-8cbcbacd3905/setup-container/0.log" Nov 24 11:35:25 crc kubenswrapper[4944]: I1124 11:35:25.796939 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_15ce2947-cdd4-4b26-99c1-8cbcbacd3905/setup-container/0.log" Nov 24 11:35:25 crc kubenswrapper[4944]: I1124 11:35:25.809904 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7d917891-d117-4058-81bd-e582da62b26e/setup-container/0.log" Nov 24 11:35:25 crc kubenswrapper[4944]: I1124 11:35:25.831801 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_15ce2947-cdd4-4b26-99c1-8cbcbacd3905/rabbitmq/0.log" Nov 24 11:35:26 crc kubenswrapper[4944]: I1124 11:35:26.091427 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7d917891-d117-4058-81bd-e582da62b26e/setup-container/0.log" Nov 24 11:35:26 crc kubenswrapper[4944]: I1124 11:35:26.110463 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-c8btv_f69ee93c-d8e0-4c43-8597-21050c074c71/reboot-os-openstack-openstack-cell1/0.log" Nov 24 11:35:26 crc kubenswrapper[4944]: I1124 11:35:26.134734 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_7d917891-d117-4058-81bd-e582da62b26e/rabbitmq/0.log" Nov 24 11:35:26 crc kubenswrapper[4944]: I1124 11:35:26.410202 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-h68cc_e3efa51f-a625-4117-85ff-1eb6f5016b3a/ssh-known-hosts-openstack/0.log" Nov 24 11:35:26 crc kubenswrapper[4944]: I1124 11:35:26.576079 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fd7272f9-a6a2-47f1-8d00-f618476ad5b5/memcached/0.log" Nov 24 11:35:26 crc kubenswrapper[4944]: I1124 11:35:26.618764 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-tdz6j_85c42bbf-4cef-46a7-9935-108aabab6020/run-os-openstack-openstack-cell1/0.log" Nov 24 11:35:26 crc kubenswrapper[4944]: I1124 11:35:26.900693 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-299ff_9a9fdd29-f803-41a5-96eb-dc882bedc95a/telemetry-openstack-openstack-cell1/0.log" Nov 24 11:35:26 crc kubenswrapper[4944]: I1124 11:35:26.902848 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-k99fs_c6c0314e-cc78-4144-a276-5badbc29bb53/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Nov 24 11:35:27 crc kubenswrapper[4944]: I1124 11:35:27.091771 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-bm2rd_bd996416-36cc-46b8-8982-78c11da1243c/validate-network-openstack-openstack-cell1/0.log" Nov 24 11:35:48 crc kubenswrapper[4944]: I1124 11:35:48.047827 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-qvqs7_f2b53031-1259-46ff-a42d-f65547877116/kube-rbac-proxy/0.log" Nov 24 11:35:48 crc kubenswrapper[4944]: I1124 11:35:48.145721 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-qvqs7_f2b53031-1259-46ff-a42d-f65547877116/manager/0.log" Nov 24 11:35:48 crc kubenswrapper[4944]: I1124 11:35:48.374304 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-cpf2g_2640bc37-2c54-4e38-ae79-a1d5cc08af65/manager/0.log" Nov 24 11:35:48 crc kubenswrapper[4944]: I1124 11:35:48.420317 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-cpf2g_2640bc37-2c54-4e38-ae79-a1d5cc08af65/kube-rbac-proxy/0.log" Nov 24 11:35:48 crc kubenswrapper[4944]: I1124 11:35:48.571965 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-864hg_28277a17-6dab-4f8e-b765-46fefba90b01/kube-rbac-proxy/0.log" Nov 24 11:35:48 crc kubenswrapper[4944]: I1124 11:35:48.614484 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-864hg_28277a17-6dab-4f8e-b765-46fefba90b01/manager/0.log" Nov 24 11:35:48 crc kubenswrapper[4944]: I1124 11:35:48.754611 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm_4ed11819-58d0-45ab-8fd6-3930861c2bc9/util/0.log" Nov 24 11:35:48 crc kubenswrapper[4944]: I1124 11:35:48.900776 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm_4ed11819-58d0-45ab-8fd6-3930861c2bc9/pull/0.log" Nov 24 11:35:48 crc kubenswrapper[4944]: I1124 11:35:48.917396 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm_4ed11819-58d0-45ab-8fd6-3930861c2bc9/util/0.log" Nov 24 11:35:48 crc kubenswrapper[4944]: I1124 11:35:48.931494 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm_4ed11819-58d0-45ab-8fd6-3930861c2bc9/pull/0.log" Nov 24 11:35:49 crc kubenswrapper[4944]: I1124 11:35:49.189333 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm_4ed11819-58d0-45ab-8fd6-3930861c2bc9/pull/0.log" Nov 24 11:35:49 crc kubenswrapper[4944]: I1124 11:35:49.192430 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm_4ed11819-58d0-45ab-8fd6-3930861c2bc9/util/0.log" Nov 24 11:35:49 crc kubenswrapper[4944]: I1124 11:35:49.195791 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eb43bae5e84b248e1aa63efc2800ff09efd3ed0938ade6192596eaf85cb7bhm_4ed11819-58d0-45ab-8fd6-3930861c2bc9/extract/0.log" Nov 24 11:35:49 crc kubenswrapper[4944]: I1124 11:35:49.401482 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-8b2jn_916de258-fc43-4fcc-bd66-9e24e3216ac8/kube-rbac-proxy/0.log" Nov 24 11:35:49 crc kubenswrapper[4944]: I1124 11:35:49.490861 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-pq27k_265e80c3-fc9b-43f0-a8f8-6f3a4562a9ca/kube-rbac-proxy/0.log" Nov 24 11:35:49 crc kubenswrapper[4944]: I1124 11:35:49.588167 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-8b2jn_916de258-fc43-4fcc-bd66-9e24e3216ac8/manager/0.log" Nov 24 11:35:49 crc kubenswrapper[4944]: I1124 11:35:49.731573 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-pq27k_265e80c3-fc9b-43f0-a8f8-6f3a4562a9ca/manager/0.log" Nov 24 11:35:49 crc kubenswrapper[4944]: I1124 11:35:49.770491 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-bnxrk_655196aa-961a-4b69-8fad-6975c0ce44f2/kube-rbac-proxy/0.log" Nov 24 11:35:49 crc kubenswrapper[4944]: I1124 11:35:49.840310 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-bnxrk_655196aa-961a-4b69-8fad-6975c0ce44f2/manager/0.log" Nov 24 11:35:49 crc kubenswrapper[4944]: I1124 11:35:49.979984 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-gvzfc_26e6c9ac-040e-4e10-97ca-6000969950f3/kube-rbac-proxy/0.log" Nov 24 11:35:50 crc kubenswrapper[4944]: I1124 11:35:50.267610 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-jq6w6_f5f1851e-a78f-458f-9c7a-fa271cfa4283/manager/0.log" Nov 24 11:35:50 crc kubenswrapper[4944]: I1124 11:35:50.283908 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-jq6w6_f5f1851e-a78f-458f-9c7a-fa271cfa4283/kube-rbac-proxy/0.log" Nov 24 11:35:50 crc kubenswrapper[4944]: I1124 11:35:50.323327 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-gvzfc_26e6c9ac-040e-4e10-97ca-6000969950f3/manager/0.log" Nov 24 11:35:50 crc kubenswrapper[4944]: I1124 11:35:50.502653 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-wgm4r_ec7ed47f-2c2c-4401-ac0a-0510d8649f01/kube-rbac-proxy/0.log" Nov 24 11:35:50 crc kubenswrapper[4944]: I1124 11:35:50.670934 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-wgm4r_ec7ed47f-2c2c-4401-ac0a-0510d8649f01/manager/0.log" Nov 24 11:35:50 crc kubenswrapper[4944]: I1124 11:35:50.723151 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-d5mp6_f62b2a83-584b-4f8a-aa62-c1e7499100bd/kube-rbac-proxy/0.log" Nov 24 11:35:50 crc kubenswrapper[4944]: I1124 11:35:50.844637 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-d5mp6_f62b2a83-584b-4f8a-aa62-c1e7499100bd/manager/0.log" Nov 24 11:35:50 crc kubenswrapper[4944]: I1124 11:35:50.977935 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-wjthz_fe8abc9f-e87f-4417-905f-44bb0b432a59/manager/0.log" Nov 24 11:35:50 crc kubenswrapper[4944]: I1124 11:35:50.985474 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-wjthz_fe8abc9f-e87f-4417-905f-44bb0b432a59/kube-rbac-proxy/0.log" Nov 24 11:35:51 crc kubenswrapper[4944]: I1124 11:35:51.187400 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-td8rp_3e0e56a5-c24a-4dea-b478-d742409c80b4/kube-rbac-proxy/0.log" Nov 24 11:35:51 crc kubenswrapper[4944]: I1124 11:35:51.279113 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-td8rp_3e0e56a5-c24a-4dea-b478-d742409c80b4/manager/0.log" Nov 24 11:35:51 crc kubenswrapper[4944]: I1124 11:35:51.432660 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-p4tcl_163609af-d6a1-48b2-a5cb-e844edff0ef4/kube-rbac-proxy/0.log" Nov 24 11:35:51 crc kubenswrapper[4944]: I1124 11:35:51.599425 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-gcggn_c702883d-7a93-4ede-a1dc-b2becaf94999/kube-rbac-proxy/0.log" Nov 24 11:35:51 crc kubenswrapper[4944]: I1124 11:35:51.608276 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-p4tcl_163609af-d6a1-48b2-a5cb-e844edff0ef4/manager/0.log" Nov 24 11:35:51 crc kubenswrapper[4944]: I1124 11:35:51.716662 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-gcggn_c702883d-7a93-4ede-a1dc-b2becaf94999/manager/0.log" Nov 24 11:35:51 crc kubenswrapper[4944]: I1124 11:35:51.909474 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf_42f09610-472c-44b1-85d4-55fa62d2c2c1/manager/0.log" Nov 24 11:35:51 crc kubenswrapper[4944]: I1124 11:35:51.961311 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-l8dmf_42f09610-472c-44b1-85d4-55fa62d2c2c1/kube-rbac-proxy/0.log" Nov 24 11:35:52 crc kubenswrapper[4944]: I1124 11:35:52.018013 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f88c7d9f9-wwjhr_f07b6604-fc2f-4507-a52c-c69c58fc4cee/kube-rbac-proxy/0.log" Nov 24 11:35:52 crc kubenswrapper[4944]: I1124 11:35:52.191652 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-56cb4fc9f6-t2lj4_35d7929f-4374-4c32-826a-f0dc8df3d8ff/kube-rbac-proxy/0.log" Nov 24 11:35:52 crc kubenswrapper[4944]: I1124 11:35:52.360466 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-56cb4fc9f6-t2lj4_35d7929f-4374-4c32-826a-f0dc8df3d8ff/operator/0.log" Nov 24 11:35:52 crc kubenswrapper[4944]: I1124 11:35:52.408863 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zgw8d_f2f1a239-2357-40b5-bf93-00ee35de0069/registry-server/0.log" Nov 24 11:35:52 crc kubenswrapper[4944]: I1124 11:35:52.496333 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-xblqz_6f6767fe-2e4b-4b7a-b92c-8007220e1f3e/kube-rbac-proxy/0.log" Nov 24 11:35:52 crc kubenswrapper[4944]: I1124 11:35:52.748871 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-dvc87_64f7e2cc-d596-4270-863c-0fa67797d789/manager/0.log" Nov 24 11:35:52 crc kubenswrapper[4944]: I1124 11:35:52.804546 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-xblqz_6f6767fe-2e4b-4b7a-b92c-8007220e1f3e/manager/0.log" Nov 24 11:35:53 crc kubenswrapper[4944]: I1124 11:35:53.053907 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-dvc87_64f7e2cc-d596-4270-863c-0fa67797d789/kube-rbac-proxy/0.log" Nov 24 11:35:53 crc kubenswrapper[4944]: I1124 11:35:53.263070 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-7wd7z_18a65aee-1bdc-4d96-8c52-711fb8cf5476/operator/0.log" Nov 24 11:35:53 crc kubenswrapper[4944]: I1124 11:35:53.301523 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-dlfx8_fed38002-b5d5-4bc0-9fde-5cb9e04c5542/kube-rbac-proxy/0.log" Nov 24 11:35:53 crc kubenswrapper[4944]: I1124 11:35:53.533329 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-dlfx8_fed38002-b5d5-4bc0-9fde-5cb9e04c5542/manager/0.log" Nov 24 11:35:53 crc kubenswrapper[4944]: I1124 11:35:53.548137 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:35:53 crc kubenswrapper[4944]: I1124 11:35:53.548226 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:35:53 crc kubenswrapper[4944]: I1124 11:35:53.626004 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-2h8kk_b4532ccb-d13d-4cd8-9ccb-240cb2d84195/kube-rbac-proxy/0.log" Nov 24 11:35:53 crc kubenswrapper[4944]: I1124 11:35:53.814844 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-2h8kk_b4532ccb-d13d-4cd8-9ccb-240cb2d84195/manager/0.log" Nov 24 11:35:53 crc kubenswrapper[4944]: I1124 11:35:53.857440 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-j5gtk_63a16eba-7d99-41f4-9c77-83a8497c4210/kube-rbac-proxy/0.log" Nov 24 11:35:53 crc kubenswrapper[4944]: I1124 11:35:53.911701 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-j5gtk_63a16eba-7d99-41f4-9c77-83a8497c4210/manager/0.log" Nov 24 11:35:54 crc kubenswrapper[4944]: I1124 11:35:54.117493 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-54vjr_82fc1284-da3b-4620-ae81-64535b5dde90/manager/0.log" Nov 24 11:35:54 crc kubenswrapper[4944]: I1124 11:35:54.168029 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-54vjr_82fc1284-da3b-4620-ae81-64535b5dde90/kube-rbac-proxy/0.log" Nov 24 11:35:54 crc kubenswrapper[4944]: I1124 11:35:54.315727 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f88c7d9f9-wwjhr_f07b6604-fc2f-4507-a52c-c69c58fc4cee/manager/0.log" Nov 24 11:36:11 crc kubenswrapper[4944]: I1124 11:36:11.484548 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jz9hv_22e1dda2-1c29-4f22-97c3-71c6aa588b47/control-plane-machine-set-operator/0.log" Nov 24 11:36:11 crc kubenswrapper[4944]: I1124 11:36:11.703716 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gg465_e6d222cd-e154-421b-afb4-dbad33103c9a/machine-api-operator/0.log" Nov 24 11:36:11 crc kubenswrapper[4944]: I1124 11:36:11.708970 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gg465_e6d222cd-e154-421b-afb4-dbad33103c9a/kube-rbac-proxy/0.log" Nov 24 11:36:23 crc kubenswrapper[4944]: I1124 11:36:23.247189 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-p7czr_8bc6288c-a8c0-4615-9ff5-1c7839c75a77/cert-manager-controller/0.log" Nov 24 11:36:23 crc kubenswrapper[4944]: I1124 11:36:23.448613 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-7gnzh_9aebbf05-0710-4819-97eb-cfb14cade6b3/cert-manager-webhook/0.log" Nov 24 11:36:23 crc kubenswrapper[4944]: I1124 11:36:23.513716 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-jt2wv_fd28b1a5-a8ce-4fc4-af42-3b3fc169e190/cert-manager-cainjector/0.log" Nov 24 11:36:23 crc kubenswrapper[4944]: I1124 11:36:23.549121 4944 patch_prober.go:28] interesting pod/machine-config-daemon-llhq9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 11:36:23 crc kubenswrapper[4944]: I1124 11:36:23.549196 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 11:36:23 crc kubenswrapper[4944]: I1124 11:36:23.549254 4944 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" Nov 24 11:36:23 crc kubenswrapper[4944]: I1124 11:36:23.549938 4944 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c"} pod="openshift-machine-config-operator/machine-config-daemon-llhq9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 11:36:23 crc kubenswrapper[4944]: I1124 11:36:23.550003 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerName="machine-config-daemon" containerID="cri-o://4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" gracePeriod=600 Nov 24 11:36:23 crc kubenswrapper[4944]: E1124 11:36:23.673180 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:36:24 crc kubenswrapper[4944]: I1124 11:36:24.237959 4944 generic.go:334] "Generic (PLEG): container finished" podID="e4e5c60c-e37d-43c8-860e-39ef930adc16" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" exitCode=0 Nov 24 11:36:24 crc kubenswrapper[4944]: I1124 11:36:24.238028 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerDied","Data":"4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c"} Nov 24 11:36:24 crc kubenswrapper[4944]: I1124 11:36:24.238320 4944 scope.go:117] "RemoveContainer" containerID="e2aa81f7eceb76b67b33340bde645da1f586279fcd8142d56e3e89877f97e54b" Nov 24 11:36:24 crc kubenswrapper[4944]: I1124 11:36:24.238955 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:36:24 crc kubenswrapper[4944]: E1124 11:36:24.239223 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:36:35 crc kubenswrapper[4944]: I1124 11:36:35.345747 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-74c8z_1a7b278b-3ae9-4b09-b586-875bb2cac1d6/nmstate-console-plugin/0.log" Nov 24 11:36:35 crc kubenswrapper[4944]: I1124 11:36:35.517530 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-vjtqb_6a5e82c5-f9a6-45ea-b831-36c6573b974e/nmstate-handler/0.log" Nov 24 11:36:35 crc kubenswrapper[4944]: I1124 11:36:35.559760 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-pqwqz_299a592e-b90c-436a-9aa4-c2b949cb0e85/kube-rbac-proxy/0.log" Nov 24 11:36:35 crc kubenswrapper[4944]: I1124 11:36:35.664001 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-pqwqz_299a592e-b90c-436a-9aa4-c2b949cb0e85/nmstate-metrics/0.log" Nov 24 11:36:35 crc kubenswrapper[4944]: I1124 11:36:35.772720 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-hzrsn_3791dd89-b8cd-4a94-9542-3929fe9a9872/nmstate-operator/0.log" Nov 24 11:36:35 crc kubenswrapper[4944]: I1124 11:36:35.856298 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-sxnnb_3f916409-7217-43dd-a0fe-859fce69591e/nmstate-webhook/0.log" Nov 24 11:36:39 crc kubenswrapper[4944]: I1124 11:36:39.277603 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:36:39 crc kubenswrapper[4944]: E1124 11:36:39.278226 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:36:49 crc kubenswrapper[4944]: I1124 11:36:49.990073 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-hfh7t_a2469128-b8f5-46f4-80a4-fe4a25aed791/kube-rbac-proxy/0.log" Nov 24 11:36:50 crc kubenswrapper[4944]: I1124 11:36:50.541632 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-hfh7t_a2469128-b8f5-46f4-80a4-fe4a25aed791/controller/0.log" Nov 24 11:36:50 crc kubenswrapper[4944]: I1124 11:36:50.865887 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-frr-files/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.128544 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-frr-files/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.143858 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-reloader/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.144967 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-reloader/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.171593 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-metrics/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.337631 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-reloader/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.367263 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-frr-files/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.376547 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-metrics/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.396182 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-metrics/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.559615 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-frr-files/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.570372 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-metrics/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.597786 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/cp-reloader/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.598287 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/controller/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.774146 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/kube-rbac-proxy-frr/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.779666 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/frr-metrics/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.786072 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/kube-rbac-proxy/0.log" Nov 24 11:36:51 crc kubenswrapper[4944]: I1124 11:36:51.972383 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/reloader/0.log" Nov 24 11:36:52 crc kubenswrapper[4944]: I1124 11:36:52.048220 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-brg9g_23ee832e-3f27-47bc-b8cb-67e0df2f5b74/frr-k8s-webhook-server/0.log" Nov 24 11:36:52 crc kubenswrapper[4944]: I1124 11:36:52.277888 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6c8c4f86d5-h8tbj_27694069-e435-46a6-9fcc-b848a4a56fde/manager/0.log" Nov 24 11:36:52 crc kubenswrapper[4944]: I1124 11:36:52.764702 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-66c67cfc8f-mnxqx_4dc5d368-9a0c-418e-981e-bde67f08e687/webhook-server/0.log" Nov 24 11:36:52 crc kubenswrapper[4944]: I1124 11:36:52.837726 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nl6qg_169a7266-9eeb-4984-86d9-6b49c647a622/kube-rbac-proxy/0.log" Nov 24 11:36:53 crc kubenswrapper[4944]: I1124 11:36:53.838397 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nl6qg_169a7266-9eeb-4984-86d9-6b49c647a622/speaker/0.log" Nov 24 11:36:54 crc kubenswrapper[4944]: I1124 11:36:54.280385 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:36:54 crc kubenswrapper[4944]: E1124 11:36:54.280851 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:36:55 crc kubenswrapper[4944]: I1124 11:36:55.178202 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbcpf_61a2baaa-dc80-42dc-b458-0e9ffb5d0e77/frr/0.log" Nov 24 11:37:06 crc kubenswrapper[4944]: I1124 11:37:06.698819 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46_a4e4656e-3009-4e6f-a478-c8cbefc40fba/util/0.log" Nov 24 11:37:06 crc kubenswrapper[4944]: I1124 11:37:06.900879 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46_a4e4656e-3009-4e6f-a478-c8cbefc40fba/util/0.log" Nov 24 11:37:06 crc kubenswrapper[4944]: I1124 11:37:06.926395 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46_a4e4656e-3009-4e6f-a478-c8cbefc40fba/pull/0.log" Nov 24 11:37:06 crc kubenswrapper[4944]: I1124 11:37:06.937487 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46_a4e4656e-3009-4e6f-a478-c8cbefc40fba/pull/0.log" Nov 24 11:37:07 crc kubenswrapper[4944]: I1124 11:37:07.128828 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46_a4e4656e-3009-4e6f-a478-c8cbefc40fba/util/0.log" Nov 24 11:37:07 crc kubenswrapper[4944]: I1124 11:37:07.129540 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46_a4e4656e-3009-4e6f-a478-c8cbefc40fba/extract/0.log" Nov 24 11:37:07 crc kubenswrapper[4944]: I1124 11:37:07.203476 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931astj46_a4e4656e-3009-4e6f-a478-c8cbefc40fba/pull/0.log" Nov 24 11:37:07 crc kubenswrapper[4944]: I1124 11:37:07.301669 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_66b08e16-cf20-4ebb-a6b0-e33bd5947b75/util/0.log" Nov 24 11:37:07 crc kubenswrapper[4944]: I1124 11:37:07.458705 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_66b08e16-cf20-4ebb-a6b0-e33bd5947b75/pull/0.log" Nov 24 11:37:07 crc kubenswrapper[4944]: I1124 11:37:07.484243 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_66b08e16-cf20-4ebb-a6b0-e33bd5947b75/util/0.log" Nov 24 11:37:07 crc kubenswrapper[4944]: I1124 11:37:07.505240 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_66b08e16-cf20-4ebb-a6b0-e33bd5947b75/pull/0.log" Nov 24 11:37:07 crc kubenswrapper[4944]: I1124 11:37:07.670979 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_66b08e16-cf20-4ebb-a6b0-e33bd5947b75/extract/0.log" Nov 24 11:37:07 crc kubenswrapper[4944]: I1124 11:37:07.708419 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_66b08e16-cf20-4ebb-a6b0-e33bd5947b75/pull/0.log" Nov 24 11:37:07 crc kubenswrapper[4944]: I1124 11:37:07.720663 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e5ll49_66b08e16-cf20-4ebb-a6b0-e33bd5947b75/util/0.log" Nov 24 11:37:07 crc kubenswrapper[4944]: I1124 11:37:07.880387 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds_181b11f3-6b16-4cfc-962e-48687b6823c1/util/0.log" Nov 24 11:37:08 crc kubenswrapper[4944]: I1124 11:37:08.058834 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds_181b11f3-6b16-4cfc-962e-48687b6823c1/util/0.log" Nov 24 11:37:08 crc kubenswrapper[4944]: I1124 11:37:08.078018 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds_181b11f3-6b16-4cfc-962e-48687b6823c1/pull/0.log" Nov 24 11:37:08 crc kubenswrapper[4944]: I1124 11:37:08.084096 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds_181b11f3-6b16-4cfc-962e-48687b6823c1/pull/0.log" Nov 24 11:37:08 crc kubenswrapper[4944]: I1124 11:37:08.240515 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds_181b11f3-6b16-4cfc-962e-48687b6823c1/pull/0.log" Nov 24 11:37:08 crc kubenswrapper[4944]: I1124 11:37:08.241575 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds_181b11f3-6b16-4cfc-962e-48687b6823c1/util/0.log" Nov 24 11:37:08 crc kubenswrapper[4944]: I1124 11:37:08.277072 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:37:08 crc kubenswrapper[4944]: E1124 11:37:08.277382 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:37:08 crc kubenswrapper[4944]: I1124 11:37:08.279177 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210rb6ds_181b11f3-6b16-4cfc-962e-48687b6823c1/extract/0.log" Nov 24 11:37:08 crc kubenswrapper[4944]: I1124 11:37:08.419714 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-wwf7x_6b0535ee-607f-481f-a60b-40909235fa96/extract-utilities/0.log" Nov 24 11:37:08 crc kubenswrapper[4944]: I1124 11:37:08.854868 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-wwf7x_6b0535ee-607f-481f-a60b-40909235fa96/extract-utilities/0.log" Nov 24 11:37:08 crc kubenswrapper[4944]: I1124 11:37:08.875891 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-wwf7x_6b0535ee-607f-481f-a60b-40909235fa96/extract-content/0.log" Nov 24 11:37:08 crc kubenswrapper[4944]: I1124 11:37:08.901989 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-wwf7x_6b0535ee-607f-481f-a60b-40909235fa96/extract-content/0.log" Nov 24 11:37:09 crc kubenswrapper[4944]: I1124 11:37:09.033155 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-wwf7x_6b0535ee-607f-481f-a60b-40909235fa96/extract-utilities/0.log" Nov 24 11:37:09 crc kubenswrapper[4944]: I1124 11:37:09.074075 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-wwf7x_6b0535ee-607f-481f-a60b-40909235fa96/extract-content/0.log" Nov 24 11:37:09 crc kubenswrapper[4944]: I1124 11:37:09.376288 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d962_c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e/extract-utilities/0.log" Nov 24 11:37:09 crc kubenswrapper[4944]: I1124 11:37:09.583317 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d962_c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e/extract-content/0.log" Nov 24 11:37:09 crc kubenswrapper[4944]: I1124 11:37:09.598133 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d962_c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e/extract-utilities/0.log" Nov 24 11:37:09 crc kubenswrapper[4944]: I1124 11:37:09.669453 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d962_c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e/extract-content/0.log" Nov 24 11:37:09 crc kubenswrapper[4944]: I1124 11:37:09.857750 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d962_c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e/extract-utilities/0.log" Nov 24 11:37:09 crc kubenswrapper[4944]: I1124 11:37:09.894952 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d962_c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e/extract-content/0.log" Nov 24 11:37:10 crc kubenswrapper[4944]: I1124 11:37:10.148167 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7_3fbed182-a3a3-46f3-b6af-6383550b36e1/util/0.log" Nov 24 11:37:10 crc kubenswrapper[4944]: I1124 11:37:10.426193 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7_3fbed182-a3a3-46f3-b6af-6383550b36e1/pull/0.log" Nov 24 11:37:10 crc kubenswrapper[4944]: I1124 11:37:10.435716 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7_3fbed182-a3a3-46f3-b6af-6383550b36e1/util/0.log" Nov 24 11:37:10 crc kubenswrapper[4944]: I1124 11:37:10.614630 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-wwf7x_6b0535ee-607f-481f-a60b-40909235fa96/registry-server/0.log" Nov 24 11:37:10 crc kubenswrapper[4944]: I1124 11:37:10.639484 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7_3fbed182-a3a3-46f3-b6af-6383550b36e1/pull/0.log" Nov 24 11:37:10 crc kubenswrapper[4944]: I1124 11:37:10.791098 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7_3fbed182-a3a3-46f3-b6af-6383550b36e1/util/0.log" Nov 24 11:37:10 crc kubenswrapper[4944]: I1124 11:37:10.903741 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7_3fbed182-a3a3-46f3-b6af-6383550b36e1/pull/0.log" Nov 24 11:37:10 crc kubenswrapper[4944]: I1124 11:37:10.923421 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6txmd7_3fbed182-a3a3-46f3-b6af-6383550b36e1/extract/0.log" Nov 24 11:37:11 crc kubenswrapper[4944]: I1124 11:37:11.111444 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wkfq7_2efbcb2c-3e38-4eda-bf92-9805f287bbb3/marketplace-operator/0.log" Nov 24 11:37:11 crc kubenswrapper[4944]: I1124 11:37:11.225848 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czxzn_2ffe67e3-61b1-401f-af45-af86b8206f85/extract-utilities/0.log" Nov 24 11:37:11 crc kubenswrapper[4944]: I1124 11:37:11.446415 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czxzn_2ffe67e3-61b1-401f-af45-af86b8206f85/extract-utilities/0.log" Nov 24 11:37:11 crc kubenswrapper[4944]: I1124 11:37:11.496019 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czxzn_2ffe67e3-61b1-401f-af45-af86b8206f85/extract-content/0.log" Nov 24 11:37:11 crc kubenswrapper[4944]: I1124 11:37:11.498229 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czxzn_2ffe67e3-61b1-401f-af45-af86b8206f85/extract-content/0.log" Nov 24 11:37:11 crc kubenswrapper[4944]: I1124 11:37:11.692489 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czxzn_2ffe67e3-61b1-401f-af45-af86b8206f85/extract-content/0.log" Nov 24 11:37:11 crc kubenswrapper[4944]: I1124 11:37:11.738746 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czxzn_2ffe67e3-61b1-401f-af45-af86b8206f85/extract-utilities/0.log" Nov 24 11:37:11 crc kubenswrapper[4944]: I1124 11:37:11.785704 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4d962_c8b3137f-dfa3-452a-a88e-4b08b6cfaa7e/registry-server/0.log" Nov 24 11:37:12 crc kubenswrapper[4944]: I1124 11:37:12.045352 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rfsn8_2b33ed49-adf4-4398-8723-e53ef109cc70/extract-utilities/0.log" Nov 24 11:37:12 crc kubenswrapper[4944]: I1124 11:37:12.063431 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czxzn_2ffe67e3-61b1-401f-af45-af86b8206f85/registry-server/0.log" Nov 24 11:37:12 crc kubenswrapper[4944]: I1124 11:37:12.123294 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rfsn8_2b33ed49-adf4-4398-8723-e53ef109cc70/extract-utilities/0.log" Nov 24 11:37:12 crc kubenswrapper[4944]: I1124 11:37:12.190862 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rfsn8_2b33ed49-adf4-4398-8723-e53ef109cc70/extract-content/0.log" Nov 24 11:37:12 crc kubenswrapper[4944]: I1124 11:37:12.203119 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rfsn8_2b33ed49-adf4-4398-8723-e53ef109cc70/extract-content/0.log" Nov 24 11:37:12 crc kubenswrapper[4944]: I1124 11:37:12.321836 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rfsn8_2b33ed49-adf4-4398-8723-e53ef109cc70/extract-utilities/0.log" Nov 24 11:37:12 crc kubenswrapper[4944]: I1124 11:37:12.350123 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rfsn8_2b33ed49-adf4-4398-8723-e53ef109cc70/extract-content/0.log" Nov 24 11:37:13 crc kubenswrapper[4944]: I1124 11:37:13.282641 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rfsn8_2b33ed49-adf4-4398-8723-e53ef109cc70/registry-server/0.log" Nov 24 11:37:21 crc kubenswrapper[4944]: I1124 11:37:21.277661 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:37:21 crc kubenswrapper[4944]: E1124 11:37:21.278403 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:37:25 crc kubenswrapper[4944]: I1124 11:37:25.151901 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-hsntg_abaf8d9b-a218-4253-8f20-eacd200299f3/prometheus-operator/0.log" Nov 24 11:37:25 crc kubenswrapper[4944]: I1124 11:37:25.311177 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-67dbf597d7-2pwfp_8cb28caf-699f-4cb4-8c4b-3967a0c5a275/prometheus-operator-admission-webhook/0.log" Nov 24 11:37:25 crc kubenswrapper[4944]: I1124 11:37:25.327266 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-67dbf597d7-8nn84_b1c3c4c6-4cf3-4653-8ed0-352ceb700925/prometheus-operator-admission-webhook/0.log" Nov 24 11:37:25 crc kubenswrapper[4944]: I1124 11:37:25.508979 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-vrrsh_b14f0425-11fb-4ed0-9599-0c08bb518a85/operator/0.log" Nov 24 11:37:25 crc kubenswrapper[4944]: I1124 11:37:25.548172 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-t7fc7_571e96c3-7ec5-4d59-8666-c9427e3d2612/perses-operator/0.log" Nov 24 11:37:34 crc kubenswrapper[4944]: I1124 11:37:34.281525 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:37:34 crc kubenswrapper[4944]: E1124 11:37:34.282264 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:37:46 crc kubenswrapper[4944]: I1124 11:37:46.277651 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:37:46 crc kubenswrapper[4944]: E1124 11:37:46.278369 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:37:57 crc kubenswrapper[4944]: I1124 11:37:57.276951 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:37:57 crc kubenswrapper[4944]: E1124 11:37:57.277949 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:38:10 crc kubenswrapper[4944]: I1124 11:38:10.287355 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:38:10 crc kubenswrapper[4944]: E1124 11:38:10.289791 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:38:25 crc kubenswrapper[4944]: I1124 11:38:25.277359 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:38:25 crc kubenswrapper[4944]: E1124 11:38:25.278212 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:38:39 crc kubenswrapper[4944]: I1124 11:38:39.277581 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:38:39 crc kubenswrapper[4944]: E1124 11:38:39.287185 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:38:51 crc kubenswrapper[4944]: I1124 11:38:51.278133 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:38:51 crc kubenswrapper[4944]: E1124 11:38:51.279240 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:39:06 crc kubenswrapper[4944]: I1124 11:39:06.277785 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:39:06 crc kubenswrapper[4944]: E1124 11:39:06.278648 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:39:19 crc kubenswrapper[4944]: I1124 11:39:19.276910 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:39:19 crc kubenswrapper[4944]: E1124 11:39:19.278715 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:39:32 crc kubenswrapper[4944]: I1124 11:39:32.277684 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:39:32 crc kubenswrapper[4944]: E1124 11:39:32.278443 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:39:35 crc kubenswrapper[4944]: I1124 11:39:35.158008 4944 generic.go:334] "Generic (PLEG): container finished" podID="e39efbba-63ef-4598-98ca-e88d7ca05ce7" containerID="ca05f81c5ee2cf26da8b2faca63e17d62ed1a742f3b14e143a906716bafc2fda" exitCode=0 Nov 24 11:39:35 crc kubenswrapper[4944]: I1124 11:39:35.158205 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gqft2/must-gather-6kf28" event={"ID":"e39efbba-63ef-4598-98ca-e88d7ca05ce7","Type":"ContainerDied","Data":"ca05f81c5ee2cf26da8b2faca63e17d62ed1a742f3b14e143a906716bafc2fda"} Nov 24 11:39:35 crc kubenswrapper[4944]: I1124 11:39:35.159699 4944 scope.go:117] "RemoveContainer" containerID="ca05f81c5ee2cf26da8b2faca63e17d62ed1a742f3b14e143a906716bafc2fda" Nov 24 11:39:35 crc kubenswrapper[4944]: I1124 11:39:35.229011 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gqft2_must-gather-6kf28_e39efbba-63ef-4598-98ca-e88d7ca05ce7/gather/0.log" Nov 24 11:39:42 crc kubenswrapper[4944]: I1124 11:39:42.958454 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gqft2/must-gather-6kf28"] Nov 24 11:39:42 crc kubenswrapper[4944]: I1124 11:39:42.959347 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gqft2/must-gather-6kf28" podUID="e39efbba-63ef-4598-98ca-e88d7ca05ce7" containerName="copy" containerID="cri-o://659ae4b751b5491017c6a35b523375044a9951da98f218bc5ba44933ddeaff94" gracePeriod=2 Nov 24 11:39:42 crc kubenswrapper[4944]: I1124 11:39:42.967439 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gqft2/must-gather-6kf28"] Nov 24 11:39:43 crc kubenswrapper[4944]: I1124 11:39:43.242309 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gqft2_must-gather-6kf28_e39efbba-63ef-4598-98ca-e88d7ca05ce7/copy/0.log" Nov 24 11:39:43 crc kubenswrapper[4944]: I1124 11:39:43.243200 4944 generic.go:334] "Generic (PLEG): container finished" podID="e39efbba-63ef-4598-98ca-e88d7ca05ce7" containerID="659ae4b751b5491017c6a35b523375044a9951da98f218bc5ba44933ddeaff94" exitCode=143 Nov 24 11:39:43 crc kubenswrapper[4944]: I1124 11:39:43.451684 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gqft2_must-gather-6kf28_e39efbba-63ef-4598-98ca-e88d7ca05ce7/copy/0.log" Nov 24 11:39:43 crc kubenswrapper[4944]: I1124 11:39:43.453227 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/must-gather-6kf28" Nov 24 11:39:43 crc kubenswrapper[4944]: I1124 11:39:43.545447 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e39efbba-63ef-4598-98ca-e88d7ca05ce7-must-gather-output\") pod \"e39efbba-63ef-4598-98ca-e88d7ca05ce7\" (UID: \"e39efbba-63ef-4598-98ca-e88d7ca05ce7\") " Nov 24 11:39:43 crc kubenswrapper[4944]: I1124 11:39:43.647151 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv569\" (UniqueName: \"kubernetes.io/projected/e39efbba-63ef-4598-98ca-e88d7ca05ce7-kube-api-access-fv569\") pod \"e39efbba-63ef-4598-98ca-e88d7ca05ce7\" (UID: \"e39efbba-63ef-4598-98ca-e88d7ca05ce7\") " Nov 24 11:39:43 crc kubenswrapper[4944]: I1124 11:39:43.653412 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e39efbba-63ef-4598-98ca-e88d7ca05ce7-kube-api-access-fv569" (OuterVolumeSpecName: "kube-api-access-fv569") pod "e39efbba-63ef-4598-98ca-e88d7ca05ce7" (UID: "e39efbba-63ef-4598-98ca-e88d7ca05ce7"). InnerVolumeSpecName "kube-api-access-fv569". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:39:43 crc kubenswrapper[4944]: I1124 11:39:43.725137 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e39efbba-63ef-4598-98ca-e88d7ca05ce7-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e39efbba-63ef-4598-98ca-e88d7ca05ce7" (UID: "e39efbba-63ef-4598-98ca-e88d7ca05ce7"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:39:43 crc kubenswrapper[4944]: I1124 11:39:43.750165 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv569\" (UniqueName: \"kubernetes.io/projected/e39efbba-63ef-4598-98ca-e88d7ca05ce7-kube-api-access-fv569\") on node \"crc\" DevicePath \"\"" Nov 24 11:39:43 crc kubenswrapper[4944]: I1124 11:39:43.750718 4944 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e39efbba-63ef-4598-98ca-e88d7ca05ce7-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 11:39:44 crc kubenswrapper[4944]: I1124 11:39:44.253710 4944 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gqft2_must-gather-6kf28_e39efbba-63ef-4598-98ca-e88d7ca05ce7/copy/0.log" Nov 24 11:39:44 crc kubenswrapper[4944]: I1124 11:39:44.254269 4944 scope.go:117] "RemoveContainer" containerID="659ae4b751b5491017c6a35b523375044a9951da98f218bc5ba44933ddeaff94" Nov 24 11:39:44 crc kubenswrapper[4944]: I1124 11:39:44.254316 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gqft2/must-gather-6kf28" Nov 24 11:39:44 crc kubenswrapper[4944]: I1124 11:39:44.296338 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:39:44 crc kubenswrapper[4944]: E1124 11:39:44.296544 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:39:44 crc kubenswrapper[4944]: I1124 11:39:44.302606 4944 scope.go:117] "RemoveContainer" containerID="ca05f81c5ee2cf26da8b2faca63e17d62ed1a742f3b14e143a906716bafc2fda" Nov 24 11:39:44 crc kubenswrapper[4944]: I1124 11:39:44.342502 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e39efbba-63ef-4598-98ca-e88d7ca05ce7" path="/var/lib/kubelet/pods/e39efbba-63ef-4598-98ca-e88d7ca05ce7/volumes" Nov 24 11:39:57 crc kubenswrapper[4944]: I1124 11:39:57.276913 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:39:57 crc kubenswrapper[4944]: E1124 11:39:57.277688 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:40:09 crc kubenswrapper[4944]: I1124 11:40:09.277034 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:40:09 crc kubenswrapper[4944]: E1124 11:40:09.277891 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:40:22 crc kubenswrapper[4944]: I1124 11:40:22.277366 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:40:22 crc kubenswrapper[4944]: E1124 11:40:22.278065 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:40:37 crc kubenswrapper[4944]: I1124 11:40:37.277751 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:40:37 crc kubenswrapper[4944]: E1124 11:40:37.278888 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:40:52 crc kubenswrapper[4944]: I1124 11:40:52.277468 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:40:52 crc kubenswrapper[4944]: E1124 11:40:52.278288 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:41:04 crc kubenswrapper[4944]: I1124 11:41:04.280029 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:41:04 crc kubenswrapper[4944]: E1124 11:41:04.281716 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:41:11 crc kubenswrapper[4944]: I1124 11:41:11.799605 4944 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="56ccc568-803f-4862-b9d6-f5e29d45f5b0" containerName="galera" probeResult="failure" output="command timed out" Nov 24 11:41:11 crc kubenswrapper[4944]: I1124 11:41:11.807473 4944 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="56ccc568-803f-4862-b9d6-f5e29d45f5b0" containerName="galera" probeResult="failure" output="command timed out" Nov 24 11:41:15 crc kubenswrapper[4944]: I1124 11:41:15.276732 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:41:15 crc kubenswrapper[4944]: E1124 11:41:15.277581 4944 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-llhq9_openshift-machine-config-operator(e4e5c60c-e37d-43c8-860e-39ef930adc16)\"" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" podUID="e4e5c60c-e37d-43c8-860e-39ef930adc16" Nov 24 11:41:28 crc kubenswrapper[4944]: I1124 11:41:28.277224 4944 scope.go:117] "RemoveContainer" containerID="4fbc575812ae54db2f7c3dc5103a6253ff639bd9eb666f2aa404973040388e6c" Nov 24 11:41:29 crc kubenswrapper[4944]: I1124 11:41:29.379355 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-llhq9" event={"ID":"e4e5c60c-e37d-43c8-860e-39ef930adc16","Type":"ContainerStarted","Data":"95d78657d9a28db10c11de7fe7dad4fb0e797e1761c86c9a55474bd1ef4aed18"} Nov 24 11:41:32 crc kubenswrapper[4944]: I1124 11:41:32.944636 4944 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d2bcf"] Nov 24 11:41:32 crc kubenswrapper[4944]: E1124 11:41:32.945683 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e39efbba-63ef-4598-98ca-e88d7ca05ce7" containerName="copy" Nov 24 11:41:32 crc kubenswrapper[4944]: I1124 11:41:32.945702 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e39efbba-63ef-4598-98ca-e88d7ca05ce7" containerName="copy" Nov 24 11:41:32 crc kubenswrapper[4944]: E1124 11:41:32.945727 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" containerName="registry-server" Nov 24 11:41:32 crc kubenswrapper[4944]: I1124 11:41:32.945735 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" containerName="registry-server" Nov 24 11:41:32 crc kubenswrapper[4944]: E1124 11:41:32.945749 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" containerName="extract-utilities" Nov 24 11:41:32 crc kubenswrapper[4944]: I1124 11:41:32.945758 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" containerName="extract-utilities" Nov 24 11:41:32 crc kubenswrapper[4944]: E1124 11:41:32.945818 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e39efbba-63ef-4598-98ca-e88d7ca05ce7" containerName="gather" Nov 24 11:41:32 crc kubenswrapper[4944]: I1124 11:41:32.945827 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="e39efbba-63ef-4598-98ca-e88d7ca05ce7" containerName="gather" Nov 24 11:41:32 crc kubenswrapper[4944]: E1124 11:41:32.945848 4944 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" containerName="extract-content" Nov 24 11:41:32 crc kubenswrapper[4944]: I1124 11:41:32.945855 4944 state_mem.go:107] "Deleted CPUSet assignment" podUID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" containerName="extract-content" Nov 24 11:41:32 crc kubenswrapper[4944]: I1124 11:41:32.946122 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e39efbba-63ef-4598-98ca-e88d7ca05ce7" containerName="gather" Nov 24 11:41:32 crc kubenswrapper[4944]: I1124 11:41:32.946159 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="d12c6fc4-dfe2-4633-b681-8bee73f872e3" containerName="registry-server" Nov 24 11:41:32 crc kubenswrapper[4944]: I1124 11:41:32.946178 4944 memory_manager.go:354] "RemoveStaleState removing state" podUID="e39efbba-63ef-4598-98ca-e88d7ca05ce7" containerName="copy" Nov 24 11:41:32 crc kubenswrapper[4944]: I1124 11:41:32.948505 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:32 crc kubenswrapper[4944]: I1124 11:41:32.964539 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d2bcf"] Nov 24 11:41:33 crc kubenswrapper[4944]: I1124 11:41:33.140789 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgx5h\" (UniqueName: \"kubernetes.io/projected/3864635b-b04f-41c2-8407-705c12eddd0a-kube-api-access-qgx5h\") pod \"redhat-marketplace-d2bcf\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:33 crc kubenswrapper[4944]: I1124 11:41:33.140864 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-utilities\") pod \"redhat-marketplace-d2bcf\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:33 crc kubenswrapper[4944]: I1124 11:41:33.141234 4944 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-catalog-content\") pod \"redhat-marketplace-d2bcf\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:33 crc kubenswrapper[4944]: I1124 11:41:33.244167 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-utilities\") pod \"redhat-marketplace-d2bcf\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:33 crc kubenswrapper[4944]: I1124 11:41:33.244360 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-catalog-content\") pod \"redhat-marketplace-d2bcf\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:33 crc kubenswrapper[4944]: I1124 11:41:33.244516 4944 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgx5h\" (UniqueName: \"kubernetes.io/projected/3864635b-b04f-41c2-8407-705c12eddd0a-kube-api-access-qgx5h\") pod \"redhat-marketplace-d2bcf\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:33 crc kubenswrapper[4944]: I1124 11:41:33.244625 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-utilities\") pod \"redhat-marketplace-d2bcf\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:33 crc kubenswrapper[4944]: I1124 11:41:33.244776 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-catalog-content\") pod \"redhat-marketplace-d2bcf\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:33 crc kubenswrapper[4944]: I1124 11:41:33.596308 4944 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgx5h\" (UniqueName: \"kubernetes.io/projected/3864635b-b04f-41c2-8407-705c12eddd0a-kube-api-access-qgx5h\") pod \"redhat-marketplace-d2bcf\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:33 crc kubenswrapper[4944]: I1124 11:41:33.890960 4944 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:34 crc kubenswrapper[4944]: I1124 11:41:34.413975 4944 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d2bcf"] Nov 24 11:41:35 crc kubenswrapper[4944]: I1124 11:41:35.444647 4944 generic.go:334] "Generic (PLEG): container finished" podID="3864635b-b04f-41c2-8407-705c12eddd0a" containerID="32110a577a41d3dcea4e4da3ac6a5652e36168a8f0723a15bd1d3b3760fe4b1c" exitCode=0 Nov 24 11:41:35 crc kubenswrapper[4944]: I1124 11:41:35.445152 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d2bcf" event={"ID":"3864635b-b04f-41c2-8407-705c12eddd0a","Type":"ContainerDied","Data":"32110a577a41d3dcea4e4da3ac6a5652e36168a8f0723a15bd1d3b3760fe4b1c"} Nov 24 11:41:35 crc kubenswrapper[4944]: I1124 11:41:35.445189 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d2bcf" event={"ID":"3864635b-b04f-41c2-8407-705c12eddd0a","Type":"ContainerStarted","Data":"a26c4670446ef515bc5ab831cfb14cdca2738b7fb72ad04dbb60aa822a68aeb5"} Nov 24 11:41:35 crc kubenswrapper[4944]: I1124 11:41:35.448337 4944 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 11:41:36 crc kubenswrapper[4944]: I1124 11:41:36.462321 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d2bcf" event={"ID":"3864635b-b04f-41c2-8407-705c12eddd0a","Type":"ContainerStarted","Data":"a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a"} Nov 24 11:41:37 crc kubenswrapper[4944]: I1124 11:41:37.472902 4944 generic.go:334] "Generic (PLEG): container finished" podID="3864635b-b04f-41c2-8407-705c12eddd0a" containerID="a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a" exitCode=0 Nov 24 11:41:37 crc kubenswrapper[4944]: I1124 11:41:37.473043 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d2bcf" event={"ID":"3864635b-b04f-41c2-8407-705c12eddd0a","Type":"ContainerDied","Data":"a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a"} Nov 24 11:41:38 crc kubenswrapper[4944]: I1124 11:41:38.487878 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d2bcf" event={"ID":"3864635b-b04f-41c2-8407-705c12eddd0a","Type":"ContainerStarted","Data":"2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435"} Nov 24 11:41:38 crc kubenswrapper[4944]: I1124 11:41:38.512251 4944 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d2bcf" podStartSLOduration=4.086363885 podStartE2EDuration="6.512230516s" podCreationTimestamp="2025-11-24 11:41:32 +0000 UTC" firstStartedPulling="2025-11-24 11:41:35.447779759 +0000 UTC m=+10155.982220221" lastFinishedPulling="2025-11-24 11:41:37.87364639 +0000 UTC m=+10158.408086852" observedRunningTime="2025-11-24 11:41:38.504822694 +0000 UTC m=+10159.039263156" watchObservedRunningTime="2025-11-24 11:41:38.512230516 +0000 UTC m=+10159.046670978" Nov 24 11:41:43 crc kubenswrapper[4944]: I1124 11:41:43.895171 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:43 crc kubenswrapper[4944]: I1124 11:41:43.896827 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:44 crc kubenswrapper[4944]: I1124 11:41:44.020593 4944 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:44 crc kubenswrapper[4944]: I1124 11:41:44.734663 4944 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:44 crc kubenswrapper[4944]: I1124 11:41:44.780031 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d2bcf"] Nov 24 11:41:46 crc kubenswrapper[4944]: I1124 11:41:46.558269 4944 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d2bcf" podUID="3864635b-b04f-41c2-8407-705c12eddd0a" containerName="registry-server" containerID="cri-o://2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435" gracePeriod=2 Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.059106 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.167949 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-utilities\") pod \"3864635b-b04f-41c2-8407-705c12eddd0a\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.168272 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgx5h\" (UniqueName: \"kubernetes.io/projected/3864635b-b04f-41c2-8407-705c12eddd0a-kube-api-access-qgx5h\") pod \"3864635b-b04f-41c2-8407-705c12eddd0a\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.168361 4944 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-catalog-content\") pod \"3864635b-b04f-41c2-8407-705c12eddd0a\" (UID: \"3864635b-b04f-41c2-8407-705c12eddd0a\") " Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.169427 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-utilities" (OuterVolumeSpecName: "utilities") pod "3864635b-b04f-41c2-8407-705c12eddd0a" (UID: "3864635b-b04f-41c2-8407-705c12eddd0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.174946 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3864635b-b04f-41c2-8407-705c12eddd0a-kube-api-access-qgx5h" (OuterVolumeSpecName: "kube-api-access-qgx5h") pod "3864635b-b04f-41c2-8407-705c12eddd0a" (UID: "3864635b-b04f-41c2-8407-705c12eddd0a"). InnerVolumeSpecName "kube-api-access-qgx5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.188459 4944 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3864635b-b04f-41c2-8407-705c12eddd0a" (UID: "3864635b-b04f-41c2-8407-705c12eddd0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.270817 4944 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgx5h\" (UniqueName: \"kubernetes.io/projected/3864635b-b04f-41c2-8407-705c12eddd0a-kube-api-access-qgx5h\") on node \"crc\" DevicePath \"\"" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.270853 4944 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.270863 4944 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3864635b-b04f-41c2-8407-705c12eddd0a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.569231 4944 generic.go:334] "Generic (PLEG): container finished" podID="3864635b-b04f-41c2-8407-705c12eddd0a" containerID="2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435" exitCode=0 Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.569295 4944 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d2bcf" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.569301 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d2bcf" event={"ID":"3864635b-b04f-41c2-8407-705c12eddd0a","Type":"ContainerDied","Data":"2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435"} Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.569752 4944 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d2bcf" event={"ID":"3864635b-b04f-41c2-8407-705c12eddd0a","Type":"ContainerDied","Data":"a26c4670446ef515bc5ab831cfb14cdca2738b7fb72ad04dbb60aa822a68aeb5"} Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.569776 4944 scope.go:117] "RemoveContainer" containerID="2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.594891 4944 scope.go:117] "RemoveContainer" containerID="a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.611136 4944 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d2bcf"] Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.633804 4944 scope.go:117] "RemoveContainer" containerID="32110a577a41d3dcea4e4da3ac6a5652e36168a8f0723a15bd1d3b3760fe4b1c" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.643943 4944 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d2bcf"] Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.671739 4944 scope.go:117] "RemoveContainer" containerID="2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435" Nov 24 11:41:47 crc kubenswrapper[4944]: E1124 11:41:47.672625 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435\": container with ID starting with 2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435 not found: ID does not exist" containerID="2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.672672 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435"} err="failed to get container status \"2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435\": rpc error: code = NotFound desc = could not find container \"2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435\": container with ID starting with 2ead49be45c53d91e837b9b5d1c27f6f25d401bfc19fad07b1587619a64ab435 not found: ID does not exist" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.672700 4944 scope.go:117] "RemoveContainer" containerID="a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a" Nov 24 11:41:47 crc kubenswrapper[4944]: E1124 11:41:47.672935 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a\": container with ID starting with a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a not found: ID does not exist" containerID="a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.673662 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a"} err="failed to get container status \"a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a\": rpc error: code = NotFound desc = could not find container \"a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a\": container with ID starting with a133c9aa594213af2120004c6f877f55a0cb5746235a2c85bc14abde6553783a not found: ID does not exist" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.673693 4944 scope.go:117] "RemoveContainer" containerID="32110a577a41d3dcea4e4da3ac6a5652e36168a8f0723a15bd1d3b3760fe4b1c" Nov 24 11:41:47 crc kubenswrapper[4944]: E1124 11:41:47.674825 4944 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32110a577a41d3dcea4e4da3ac6a5652e36168a8f0723a15bd1d3b3760fe4b1c\": container with ID starting with 32110a577a41d3dcea4e4da3ac6a5652e36168a8f0723a15bd1d3b3760fe4b1c not found: ID does not exist" containerID="32110a577a41d3dcea4e4da3ac6a5652e36168a8f0723a15bd1d3b3760fe4b1c" Nov 24 11:41:47 crc kubenswrapper[4944]: I1124 11:41:47.674857 4944 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32110a577a41d3dcea4e4da3ac6a5652e36168a8f0723a15bd1d3b3760fe4b1c"} err="failed to get container status \"32110a577a41d3dcea4e4da3ac6a5652e36168a8f0723a15bd1d3b3760fe4b1c\": rpc error: code = NotFound desc = could not find container \"32110a577a41d3dcea4e4da3ac6a5652e36168a8f0723a15bd1d3b3760fe4b1c\": container with ID starting with 32110a577a41d3dcea4e4da3ac6a5652e36168a8f0723a15bd1d3b3760fe4b1c not found: ID does not exist" Nov 24 11:41:48 crc kubenswrapper[4944]: I1124 11:41:48.301410 4944 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3864635b-b04f-41c2-8407-705c12eddd0a" path="/var/lib/kubelet/pods/3864635b-b04f-41c2-8407-705c12eddd0a/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111042347024443 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111042350017352 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111015722016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111015722015450 5ustar corecore